var/home/core/zuul-output/0000755000175000017500000000000015123706376014537 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015123724005015470 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005722553115123723776017724 0ustar rootrootDec 27 07:42:18 crc systemd[1]: Starting Kubernetes Kubelet... Dec 27 07:42:18 crc restorecon[4703]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:18 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 27 07:42:19 crc restorecon[4703]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 27 07:42:19 crc restorecon[4703]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 27 07:42:19 crc kubenswrapper[4934]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 27 07:42:19 crc kubenswrapper[4934]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 27 07:42:19 crc kubenswrapper[4934]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 27 07:42:19 crc kubenswrapper[4934]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 27 07:42:19 crc kubenswrapper[4934]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 27 07:42:19 crc kubenswrapper[4934]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.298357 4934 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.305793 4934 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.305837 4934 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.305846 4934 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.305857 4934 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.305866 4934 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.305875 4934 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.305883 4934 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.305895 4934 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.305904 4934 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.305913 4934 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.305921 4934 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.305929 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.305936 4934 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.305946 4934 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.305955 4934 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.305964 4934 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.305971 4934 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.305979 4934 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.305986 4934 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.305997 4934 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306006 4934 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306014 4934 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306023 4934 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306031 4934 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306040 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306047 4934 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306054 4934 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306062 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306069 4934 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306077 4934 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306111 4934 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306119 4934 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306126 4934 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306144 4934 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306153 4934 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306161 4934 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306169 4934 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306179 4934 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306189 4934 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306197 4934 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306205 4934 feature_gate.go:330] unrecognized feature gate: Example Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306213 4934 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306221 4934 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306230 4934 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306237 4934 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306246 4934 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306254 4934 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306262 4934 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306269 4934 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306277 4934 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306285 4934 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306293 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306300 4934 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306308 4934 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306316 4934 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306323 4934 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306331 4934 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306339 4934 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306346 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306353 4934 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306363 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306371 4934 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306378 4934 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306385 4934 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306393 4934 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306400 4934 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306408 4934 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306416 4934 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306424 4934 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306431 4934 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.306439 4934 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306598 4934 flags.go:64] FLAG: --address="0.0.0.0" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306615 4934 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306628 4934 flags.go:64] FLAG: --anonymous-auth="true" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306643 4934 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306655 4934 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306665 4934 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306677 4934 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306687 4934 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306697 4934 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306706 4934 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306716 4934 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306726 4934 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306735 4934 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306744 4934 flags.go:64] FLAG: --cgroup-root="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306753 4934 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306763 4934 flags.go:64] FLAG: --client-ca-file="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306772 4934 flags.go:64] FLAG: --cloud-config="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306781 4934 flags.go:64] FLAG: --cloud-provider="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306790 4934 flags.go:64] FLAG: --cluster-dns="[]" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306802 4934 flags.go:64] FLAG: --cluster-domain="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306810 4934 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306820 4934 flags.go:64] FLAG: --config-dir="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306828 4934 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306838 4934 flags.go:64] FLAG: --container-log-max-files="5" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306849 4934 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306858 4934 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306867 4934 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306876 4934 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306885 4934 flags.go:64] FLAG: --contention-profiling="false" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306894 4934 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306903 4934 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306912 4934 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306921 4934 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306932 4934 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306942 4934 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306950 4934 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306959 4934 flags.go:64] FLAG: --enable-load-reader="false" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306968 4934 flags.go:64] FLAG: --enable-server="true" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306976 4934 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306987 4934 flags.go:64] FLAG: --event-burst="100" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.306996 4934 flags.go:64] FLAG: --event-qps="50" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307005 4934 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307014 4934 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307023 4934 flags.go:64] FLAG: --eviction-hard="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307033 4934 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307043 4934 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307052 4934 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307062 4934 flags.go:64] FLAG: --eviction-soft="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307071 4934 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307107 4934 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307116 4934 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307125 4934 flags.go:64] FLAG: --experimental-mounter-path="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307134 4934 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307143 4934 flags.go:64] FLAG: --fail-swap-on="true" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307152 4934 flags.go:64] FLAG: --feature-gates="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307163 4934 flags.go:64] FLAG: --file-check-frequency="20s" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307172 4934 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307180 4934 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307190 4934 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307199 4934 flags.go:64] FLAG: --healthz-port="10248" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307208 4934 flags.go:64] FLAG: --help="false" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307216 4934 flags.go:64] FLAG: --hostname-override="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307226 4934 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307236 4934 flags.go:64] FLAG: --http-check-frequency="20s" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307244 4934 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307253 4934 flags.go:64] FLAG: --image-credential-provider-config="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307261 4934 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307270 4934 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307280 4934 flags.go:64] FLAG: --image-service-endpoint="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307288 4934 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307297 4934 flags.go:64] FLAG: --kube-api-burst="100" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307306 4934 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307315 4934 flags.go:64] FLAG: --kube-api-qps="50" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307325 4934 flags.go:64] FLAG: --kube-reserved="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307334 4934 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307343 4934 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307352 4934 flags.go:64] FLAG: --kubelet-cgroups="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307363 4934 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307372 4934 flags.go:64] FLAG: --lock-file="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307381 4934 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307390 4934 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307400 4934 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307423 4934 flags.go:64] FLAG: --log-json-split-stream="false" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307433 4934 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307442 4934 flags.go:64] FLAG: --log-text-split-stream="false" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307451 4934 flags.go:64] FLAG: --logging-format="text" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307460 4934 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307470 4934 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307478 4934 flags.go:64] FLAG: --manifest-url="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307487 4934 flags.go:64] FLAG: --manifest-url-header="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307499 4934 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307508 4934 flags.go:64] FLAG: --max-open-files="1000000" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307518 4934 flags.go:64] FLAG: --max-pods="110" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307527 4934 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307536 4934 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307545 4934 flags.go:64] FLAG: --memory-manager-policy="None" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307554 4934 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307563 4934 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307571 4934 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307580 4934 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307601 4934 flags.go:64] FLAG: --node-status-max-images="50" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307609 4934 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307618 4934 flags.go:64] FLAG: --oom-score-adj="-999" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307628 4934 flags.go:64] FLAG: --pod-cidr="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307636 4934 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307650 4934 flags.go:64] FLAG: --pod-manifest-path="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307661 4934 flags.go:64] FLAG: --pod-max-pids="-1" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307672 4934 flags.go:64] FLAG: --pods-per-core="0" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307683 4934 flags.go:64] FLAG: --port="10250" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307695 4934 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307707 4934 flags.go:64] FLAG: --provider-id="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307718 4934 flags.go:64] FLAG: --qos-reserved="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307729 4934 flags.go:64] FLAG: --read-only-port="10255" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307776 4934 flags.go:64] FLAG: --register-node="true" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307789 4934 flags.go:64] FLAG: --register-schedulable="true" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307798 4934 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307814 4934 flags.go:64] FLAG: --registry-burst="10" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307824 4934 flags.go:64] FLAG: --registry-qps="5" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307832 4934 flags.go:64] FLAG: --reserved-cpus="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307844 4934 flags.go:64] FLAG: --reserved-memory="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307858 4934 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307869 4934 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307881 4934 flags.go:64] FLAG: --rotate-certificates="false" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307891 4934 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307900 4934 flags.go:64] FLAG: --runonce="false" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307909 4934 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307918 4934 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307928 4934 flags.go:64] FLAG: --seccomp-default="false" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307937 4934 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307945 4934 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307955 4934 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307964 4934 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307973 4934 flags.go:64] FLAG: --storage-driver-password="root" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307982 4934 flags.go:64] FLAG: --storage-driver-secure="false" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.307990 4934 flags.go:64] FLAG: --storage-driver-table="stats" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.308000 4934 flags.go:64] FLAG: --storage-driver-user="root" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.308009 4934 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.308019 4934 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.308028 4934 flags.go:64] FLAG: --system-cgroups="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.308037 4934 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.308052 4934 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.308061 4934 flags.go:64] FLAG: --tls-cert-file="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.308070 4934 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.308109 4934 flags.go:64] FLAG: --tls-min-version="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.308118 4934 flags.go:64] FLAG: --tls-private-key-file="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.308127 4934 flags.go:64] FLAG: --topology-manager-policy="none" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.308136 4934 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.308145 4934 flags.go:64] FLAG: --topology-manager-scope="container" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.308154 4934 flags.go:64] FLAG: --v="2" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.308165 4934 flags.go:64] FLAG: --version="false" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.308176 4934 flags.go:64] FLAG: --vmodule="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.308187 4934 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.308197 4934 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308411 4934 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308421 4934 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308431 4934 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308440 4934 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308448 4934 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308456 4934 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308464 4934 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308472 4934 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308482 4934 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308492 4934 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308500 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308508 4934 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308516 4934 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308523 4934 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308531 4934 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308539 4934 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308546 4934 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308554 4934 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308562 4934 feature_gate.go:330] unrecognized feature gate: Example Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308570 4934 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308578 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308586 4934 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308594 4934 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308601 4934 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308609 4934 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308619 4934 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308628 4934 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308637 4934 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308645 4934 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308653 4934 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308661 4934 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308669 4934 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308677 4934 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308684 4934 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308692 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308700 4934 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308708 4934 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308716 4934 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308725 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308735 4934 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308745 4934 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308753 4934 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308762 4934 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308770 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308778 4934 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308787 4934 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308795 4934 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308802 4934 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308810 4934 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308820 4934 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308830 4934 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308840 4934 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308849 4934 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308893 4934 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308903 4934 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308912 4934 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308922 4934 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308931 4934 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308940 4934 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308949 4934 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308957 4934 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308964 4934 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308972 4934 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308980 4934 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308987 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.308995 4934 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.309003 4934 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.309010 4934 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.309018 4934 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.309025 4934 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.309033 4934 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.309055 4934 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.319886 4934 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.319950 4934 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320139 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320163 4934 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320176 4934 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320192 4934 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320203 4934 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320212 4934 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320223 4934 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320233 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320243 4934 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320252 4934 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320263 4934 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320273 4934 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320282 4934 feature_gate.go:330] unrecognized feature gate: Example Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320292 4934 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320300 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320308 4934 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320316 4934 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320324 4934 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320333 4934 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320343 4934 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320353 4934 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320363 4934 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320376 4934 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320389 4934 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320401 4934 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320412 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320423 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320435 4934 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320445 4934 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320455 4934 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320464 4934 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320472 4934 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320481 4934 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320489 4934 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320500 4934 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320509 4934 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320516 4934 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320524 4934 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320532 4934 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320540 4934 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320548 4934 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320555 4934 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320563 4934 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320571 4934 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320579 4934 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320587 4934 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320594 4934 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320602 4934 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320609 4934 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320617 4934 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320624 4934 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320632 4934 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320639 4934 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320647 4934 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320654 4934 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320662 4934 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320669 4934 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320677 4934 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320685 4934 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320693 4934 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320701 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320708 4934 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320716 4934 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320724 4934 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320732 4934 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320740 4934 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320747 4934 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320755 4934 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320762 4934 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320770 4934 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.320791 4934 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.320807 4934 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321042 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321056 4934 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321066 4934 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321075 4934 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321113 4934 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321122 4934 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321132 4934 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321142 4934 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321151 4934 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321160 4934 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321169 4934 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321177 4934 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321185 4934 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321195 4934 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321204 4934 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321212 4934 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321220 4934 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321228 4934 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321236 4934 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321244 4934 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321253 4934 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321261 4934 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321268 4934 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321276 4934 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321285 4934 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321293 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321300 4934 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321308 4934 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321315 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321323 4934 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321334 4934 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321345 4934 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321354 4934 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321363 4934 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321373 4934 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321382 4934 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321390 4934 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321398 4934 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321406 4934 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321414 4934 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321422 4934 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321430 4934 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321438 4934 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321446 4934 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321453 4934 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321461 4934 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321469 4934 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321477 4934 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321487 4934 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321499 4934 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321509 4934 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321520 4934 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321532 4934 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321544 4934 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321554 4934 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321564 4934 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321574 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321584 4934 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321594 4934 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321604 4934 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321614 4934 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321623 4934 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321633 4934 feature_gate.go:330] unrecognized feature gate: Example Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321642 4934 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321652 4934 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321661 4934 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321671 4934 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321681 4934 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321691 4934 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321701 4934 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.321739 4934 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.321758 4934 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.322130 4934 server.go:940] "Client rotation is on, will bootstrap in background" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.327043 4934 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.327263 4934 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.328237 4934 server.go:997] "Starting client certificate rotation" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.328281 4934 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.328859 4934 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-09 21:41:28.594062401 +0000 UTC Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.329079 4934 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.337263 4934 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 27 07:42:19 crc kubenswrapper[4934]: E1227 07:42:19.339902 4934 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.341073 4934 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.352076 4934 log.go:25] "Validated CRI v1 runtime API" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.368470 4934 log.go:25] "Validated CRI v1 image API" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.370117 4934 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.373775 4934 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-27-07-37-40-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.373810 4934 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.392189 4934 manager.go:217] Machine: {Timestamp:2025-12-27 07:42:19.390612996 +0000 UTC m=+0.211053630 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:e8025479-33d0-4936-ac24-c37235bb6b99 BootID:27320adc-74b9-4bb8-b209-d917a73b2290 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:68:80:f1 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:68:80:f1 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:c0:99:34 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:9b:39:a7 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:fa:bc:03 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:6a:f8:2c Speed:-1 Mtu:1496} {Name:eth10 MacAddress:76:01:06:58:92:3d Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:c2:6d:4c:4b:3a:02 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.392503 4934 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.392649 4934 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.393945 4934 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.394343 4934 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.394413 4934 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.395007 4934 topology_manager.go:138] "Creating topology manager with none policy" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.395055 4934 container_manager_linux.go:303] "Creating device plugin manager" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.395443 4934 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.395510 4934 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.395827 4934 state_mem.go:36] "Initialized new in-memory state store" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.395992 4934 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.397064 4934 kubelet.go:418] "Attempting to sync node with API server" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.397129 4934 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.397157 4934 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.397179 4934 kubelet.go:324] "Adding apiserver pod source" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.397196 4934 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.399419 4934 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.399544 4934 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 27 07:42:19 crc kubenswrapper[4934]: E1227 07:42:19.399553 4934 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.399529 4934 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Dec 27 07:42:19 crc kubenswrapper[4934]: E1227 07:42:19.399776 4934 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.400010 4934 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.400850 4934 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.401497 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.401526 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.401535 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.401544 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.401560 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.401569 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.401578 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.401592 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.401603 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.401613 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.401643 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.401652 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.402406 4934 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.402937 4934 server.go:1280] "Started kubelet" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.403479 4934 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.404156 4934 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.404961 4934 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Dec 27 07:42:19 crc systemd[1]: Started Kubernetes Kubelet. Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.405325 4934 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.405462 4934 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.405606 4934 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.405866 4934 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.405884 4934 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.405996 4934 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.408206 4934 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 08:06:07.88338237 +0000 UTC Dec 27 07:42:19 crc kubenswrapper[4934]: E1227 07:42:19.408389 4934 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.409394 4934 factory.go:55] Registering systemd factory Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.409445 4934 factory.go:221] Registration of the systemd container factory successfully Dec 27 07:42:19 crc kubenswrapper[4934]: E1227 07:42:19.409341 4934 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" interval="200ms" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.410078 4934 factory.go:153] Registering CRI-O factory Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.410190 4934 factory.go:221] Registration of the crio container factory successfully Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.410336 4934 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.410393 4934 factory.go:103] Registering Raw factory Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.410435 4934 manager.go:1196] Started watching for new ooms in manager Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.410773 4934 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Dec 27 07:42:19 crc kubenswrapper[4934]: E1227 07:42:19.410906 4934 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.413237 4934 manager.go:319] Starting recovery of all containers Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.417160 4934 server.go:460] "Adding debug handlers to kubelet server" Dec 27 07:42:19 crc kubenswrapper[4934]: E1227 07:42:19.416201 4934 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.13:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.188502a5acb0a9b9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-27 07:42:19.402906041 +0000 UTC m=+0.223346655,LastTimestamp:2025-12-27 07:42:19.402906041 +0000 UTC m=+0.223346655,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.428310 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.428429 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.428452 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.428471 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.428491 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.428509 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.428538 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.428566 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.428586 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.428612 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.428630 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.428649 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.428668 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.428767 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.428787 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.428845 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.428868 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.428893 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.428917 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.428940 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.428965 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429026 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429163 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429228 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429340 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429369 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429471 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429498 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429520 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429538 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429555 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429599 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429616 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429633 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429694 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429713 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429729 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429746 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429816 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429893 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429912 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429940 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429968 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.429987 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430018 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430056 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430127 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430205 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430225 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430243 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430259 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430278 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430325 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430372 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430393 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430436 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430482 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430501 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430518 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430535 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430554 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430572 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430599 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430645 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430664 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430682 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430726 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430813 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430851 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430869 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430888 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.430929 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.431008 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.431049 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.431346 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.431395 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.431423 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.431448 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.431465 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.431581 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.431629 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.431709 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.431736 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.431760 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.431784 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.431840 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.431864 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.431916 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.431962 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.431987 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.432012 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.432037 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.432061 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.432119 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.432148 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.432217 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.432243 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.432331 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.432378 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.432407 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.432431 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.432455 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.432481 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.432550 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.432780 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.432842 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.432864 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.432893 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.432915 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.432962 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.432984 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433033 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433140 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433190 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433209 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433260 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433297 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433343 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433360 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433401 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433442 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433458 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433477 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433493 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433509 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433526 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433543 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433585 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433604 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433652 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433729 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433830 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433867 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433885 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433903 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433941 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433958 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433974 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.433991 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434009 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434026 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434042 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434059 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434161 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434179 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434216 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434233 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434251 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434312 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434355 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434410 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434502 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434567 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434586 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434625 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434642 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434743 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434777 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434853 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434887 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434908 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434933 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434957 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.434980 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.435002 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.435019 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.435036 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.435054 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.435071 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.435118 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.435136 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.435155 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.435173 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.435191 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.435209 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.435227 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.435244 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.435260 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.435277 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.435295 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.435310 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.435329 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.435346 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.435363 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.437563 4934 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.437643 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.437673 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.437698 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.437721 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.437744 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.437763 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.437786 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.437807 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.437832 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.437852 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.437873 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.437907 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.437927 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.437947 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.437967 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.437988 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.438009 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.438028 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.438051 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.438071 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.438129 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.438149 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.438184 4934 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.438203 4934 reconstruct.go:97] "Volume reconstruction finished" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.438217 4934 reconciler.go:26] "Reconciler: start to sync state" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.455040 4934 manager.go:324] Recovery completed Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.463639 4934 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.465103 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.466004 4934 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.466071 4934 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.466164 4934 kubelet.go:2335] "Starting kubelet main sync loop" Dec 27 07:42:19 crc kubenswrapper[4934]: E1227 07:42:19.466220 4934 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.466516 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.466604 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.466617 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.467199 4934 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Dec 27 07:42:19 crc kubenswrapper[4934]: E1227 07:42:19.467322 4934 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.467792 4934 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.467823 4934 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.467852 4934 state_mem.go:36] "Initialized new in-memory state store" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.478233 4934 policy_none.go:49] "None policy: Start" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.479513 4934 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.479539 4934 state_mem.go:35] "Initializing new in-memory state store" Dec 27 07:42:19 crc kubenswrapper[4934]: E1227 07:42:19.509360 4934 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.556187 4934 manager.go:334] "Starting Device Plugin manager" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.556260 4934 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.556278 4934 server.go:79] "Starting device plugin registration server" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.556842 4934 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.556866 4934 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.557100 4934 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.557253 4934 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.557277 4934 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 27 07:42:19 crc kubenswrapper[4934]: E1227 07:42:19.565759 4934 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.566975 4934 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.567068 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.568011 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.568045 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.568057 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.568223 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.568482 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.568553 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.568900 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.568924 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.568937 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.569042 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.569188 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.569219 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.569333 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.569358 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.569368 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.569654 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.569680 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.569690 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.569829 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.570043 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.570103 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.570263 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.570302 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.570315 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.570595 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.570625 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.570635 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.570730 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.570835 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.570867 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.571463 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.571492 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.571524 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.571691 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.571710 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.571729 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.571741 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.571717 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.571784 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.571871 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.571901 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.572583 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.572608 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.572618 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:19 crc kubenswrapper[4934]: E1227 07:42:19.610775 4934 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" interval="400ms" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.640078 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.640139 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.640159 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.640178 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.640192 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.640349 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.640458 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.640529 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.640722 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.640818 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.640892 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.640946 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.640992 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.641034 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.641075 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.657883 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.659477 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.659540 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.659558 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.659594 4934 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 27 07:42:19 crc kubenswrapper[4934]: E1227 07:42:19.660229 4934 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.13:6443: connect: connection refused" node="crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.741721 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.742682 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.742748 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.742780 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.742809 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.742843 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.742878 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.742909 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.742968 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.743026 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.743109 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.743151 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.743182 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.743210 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.743238 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.742113 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.743825 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.743912 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.743957 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.744008 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.744046 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.744124 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.744059 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.744178 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.744231 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.744238 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.744314 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.744403 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.744490 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.744599 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.860682 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.862113 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.862147 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.862160 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.862186 4934 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 27 07:42:19 crc kubenswrapper[4934]: E1227 07:42:19.862581 4934 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.13:6443: connect: connection refused" node="crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.903110 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.909184 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.924328 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.936375 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-17e9030a10610360bc3c75ceda36fcd4867077057b6b5f60e8a50f1985973d94 WatchSource:0}: Error finding container 17e9030a10610360bc3c75ceda36fcd4867077057b6b5f60e8a50f1985973d94: Status 404 returned error can't find the container with id 17e9030a10610360bc3c75ceda36fcd4867077057b6b5f60e8a50f1985973d94 Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.938235 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-53994f84148b7c1fe7a91225b593cc08ed73133ca1be54d32f66cacaf2b1367d WatchSource:0}: Error finding container 53994f84148b7c1fe7a91225b593cc08ed73133ca1be54d32f66cacaf2b1367d: Status 404 returned error can't find the container with id 53994f84148b7c1fe7a91225b593cc08ed73133ca1be54d32f66cacaf2b1367d Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.940741 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.945614 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-3810eac71a0a930254e488025cbfb7a892f04634a24b3e3aae93c20ceb27234b WatchSource:0}: Error finding container 3810eac71a0a930254e488025cbfb7a892f04634a24b3e3aae93c20ceb27234b: Status 404 returned error can't find the container with id 3810eac71a0a930254e488025cbfb7a892f04634a24b3e3aae93c20ceb27234b Dec 27 07:42:19 crc kubenswrapper[4934]: I1227 07:42:19.947399 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.958985 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-6c442a7e8e0863674ad6b0599fa660d54006446c2384da43c014f2e9f3824878 WatchSource:0}: Error finding container 6c442a7e8e0863674ad6b0599fa660d54006446c2384da43c014f2e9f3824878: Status 404 returned error can't find the container with id 6c442a7e8e0863674ad6b0599fa660d54006446c2384da43c014f2e9f3824878 Dec 27 07:42:19 crc kubenswrapper[4934]: W1227 07:42:19.971417 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-0696408a701e03d298a421ede335bfcf95e54ad412d885a6c6751ac912922486 WatchSource:0}: Error finding container 0696408a701e03d298a421ede335bfcf95e54ad412d885a6c6751ac912922486: Status 404 returned error can't find the container with id 0696408a701e03d298a421ede335bfcf95e54ad412d885a6c6751ac912922486 Dec 27 07:42:20 crc kubenswrapper[4934]: E1227 07:42:20.012297 4934 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" interval="800ms" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.263718 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.265109 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.265148 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.265160 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.265184 4934 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 27 07:42:20 crc kubenswrapper[4934]: E1227 07:42:20.265666 4934 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.13:6443: connect: connection refused" node="crc" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.407076 4934 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.409378 4934 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 08:19:01.560094456 +0000 UTC Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.473131 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912"} Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.473245 4934 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912" exitCode=0 Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.473360 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.473370 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0696408a701e03d298a421ede335bfcf95e54ad412d885a6c6751ac912922486"} Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.474992 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.475036 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.475049 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.475217 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47"} Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.475253 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6c442a7e8e0863674ad6b0599fa660d54006446c2384da43c014f2e9f3824878"} Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.477049 4934 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e" exitCode=0 Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.477150 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e"} Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.477197 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3810eac71a0a930254e488025cbfb7a892f04634a24b3e3aae93c20ceb27234b"} Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.477267 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.478137 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.478159 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.478187 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.479828 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.479908 4934 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b202f5c0c838baa712ab5177b345c75f89f955286fbb832c0fa54ad0102646cd" exitCode=0 Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.479966 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b202f5c0c838baa712ab5177b345c75f89f955286fbb832c0fa54ad0102646cd"} Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.479997 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"17e9030a10610360bc3c75ceda36fcd4867077057b6b5f60e8a50f1985973d94"} Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.480139 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.480835 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.480911 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.480931 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.480915 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.481007 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.481032 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.482982 4934 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="1cdf7e91d7475938502db6b0530157177f1c38196efa5b8821d9dfec51640922" exitCode=0 Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.483018 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"1cdf7e91d7475938502db6b0530157177f1c38196efa5b8821d9dfec51640922"} Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.483041 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"53994f84148b7c1fe7a91225b593cc08ed73133ca1be54d32f66cacaf2b1367d"} Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.483133 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.483819 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.483861 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:20 crc kubenswrapper[4934]: I1227 07:42:20.483874 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:20 crc kubenswrapper[4934]: W1227 07:42:20.521538 4934 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Dec 27 07:42:20 crc kubenswrapper[4934]: E1227 07:42:20.521639 4934 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Dec 27 07:42:20 crc kubenswrapper[4934]: W1227 07:42:20.646636 4934 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Dec 27 07:42:20 crc kubenswrapper[4934]: E1227 07:42:20.646734 4934 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Dec 27 07:42:20 crc kubenswrapper[4934]: E1227 07:42:20.813121 4934 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" interval="1.6s" Dec 27 07:42:20 crc kubenswrapper[4934]: W1227 07:42:20.950874 4934 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Dec 27 07:42:20 crc kubenswrapper[4934]: E1227 07:42:20.951019 4934 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Dec 27 07:42:20 crc kubenswrapper[4934]: W1227 07:42:20.973571 4934 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Dec 27 07:42:20 crc kubenswrapper[4934]: E1227 07:42:20.973660 4934 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.065924 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.068039 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.068125 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.068135 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.068158 4934 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.410146 4934 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 01:45:18.692086164 +0000 UTC Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.441412 4934 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.486755 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3"} Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.486794 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f"} Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.486804 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781"} Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.486881 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.487586 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.487609 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.487617 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.489571 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4"} Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.489592 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8"} Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.489601 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9"} Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.489609 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787"} Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.490752 4934 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9ea9ab059166c58229e32334033a120ffde4f8b5212c50c55c97d29eab0f08c0" exitCode=0 Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.490784 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9ea9ab059166c58229e32334033a120ffde4f8b5212c50c55c97d29eab0f08c0"} Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.490848 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.497918 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.497948 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.497957 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.500163 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"1bae82234df2e64ff6d6b36e8a2aa39cf767f34d9804d5cdf2566c1efa23ad5b"} Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.500243 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.501326 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.501346 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.501353 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.503475 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"243b20c55923c82f8d7a735a9bd0656bfe13008645b5b637c9d0f3890b5890d3"} Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.503502 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e461c87813b05a23e571f459ad229874b607be5916ab31f502d476655d5f3d2b"} Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.503512 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"80370fd8607ddb05cccbe7383b94c5a4c1a874299bbbc7f665805ae5edf73ef9"} Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.503573 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.504219 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.504235 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:21 crc kubenswrapper[4934]: I1227 07:42:21.504243 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.391178 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.401698 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.410844 4934 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 03:50:41.323269596 +0000 UTC Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.510811 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a"} Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.510919 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.512395 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.512428 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.512441 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.513846 4934 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="346ccb1fac116b2783117f09752cc5c013a848dbb02b893ed052989b84af4a81" exitCode=0 Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.513945 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.513942 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"346ccb1fac116b2783117f09752cc5c013a848dbb02b893ed052989b84af4a81"} Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.514011 4934 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.514058 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.514155 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.515312 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.515342 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.515354 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.515448 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.515464 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.515484 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.515493 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.515500 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.515515 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:22 crc kubenswrapper[4934]: I1227 07:42:22.565967 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:42:23 crc kubenswrapper[4934]: I1227 07:42:23.159264 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:42:23 crc kubenswrapper[4934]: I1227 07:42:23.411983 4934 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 20:12:25.470502532 +0000 UTC Dec 27 07:42:23 crc kubenswrapper[4934]: I1227 07:42:23.519094 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"999f8cf9a8526d9f8c1f9aa47a2e86b13e4a0fbdfb872b5be86c9a72fa211c2b"} Dec 27 07:42:23 crc kubenswrapper[4934]: I1227 07:42:23.519135 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6e4686405be63521dfa03ce0d3940e52a91b7724cc41c18afd134a70f2218985"} Dec 27 07:42:23 crc kubenswrapper[4934]: I1227 07:42:23.519149 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"34214cc117a89326526885d2667b71811860562c68acc95014c51bbe7996a9af"} Dec 27 07:42:23 crc kubenswrapper[4934]: I1227 07:42:23.519203 4934 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 27 07:42:23 crc kubenswrapper[4934]: I1227 07:42:23.519228 4934 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 27 07:42:23 crc kubenswrapper[4934]: I1227 07:42:23.519251 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:23 crc kubenswrapper[4934]: I1227 07:42:23.519277 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:23 crc kubenswrapper[4934]: I1227 07:42:23.520242 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:23 crc kubenswrapper[4934]: I1227 07:42:23.520289 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:23 crc kubenswrapper[4934]: I1227 07:42:23.520306 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:23 crc kubenswrapper[4934]: I1227 07:42:23.520344 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:23 crc kubenswrapper[4934]: I1227 07:42:23.520369 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:23 crc kubenswrapper[4934]: I1227 07:42:23.520394 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.144476 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.144706 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.146208 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.146282 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.146308 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.412459 4934 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 06:09:20.765126548 +0000 UTC Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.528207 4934 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.528255 4934 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.528271 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.528322 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.528276 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.529072 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b4ef5571db679ca321aed3c4d27699e8ed04433820df8f82089d548b25860d8e"} Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.529241 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2bbcfef8d68056aba2996d89fcac8d45b1457a76a1fd9320ffd6437a0f7b5705"} Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.530154 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.530211 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.530233 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.530307 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.530307 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.530356 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.530372 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.530391 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:24 crc kubenswrapper[4934]: I1227 07:42:24.530395 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:25 crc kubenswrapper[4934]: I1227 07:42:25.413189 4934 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 05:05:53.988373602 +0000 UTC Dec 27 07:42:25 crc kubenswrapper[4934]: I1227 07:42:25.530405 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:25 crc kubenswrapper[4934]: I1227 07:42:25.531573 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:25 crc kubenswrapper[4934]: I1227 07:42:25.531613 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:25 crc kubenswrapper[4934]: I1227 07:42:25.531622 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:25 crc kubenswrapper[4934]: I1227 07:42:25.588276 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 27 07:42:25 crc kubenswrapper[4934]: I1227 07:42:25.648936 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 27 07:42:26 crc kubenswrapper[4934]: I1227 07:42:26.165304 4934 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 07:42:26 crc kubenswrapper[4934]: I1227 07:42:26.165426 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 07:42:26 crc kubenswrapper[4934]: I1227 07:42:26.353014 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:42:26 crc kubenswrapper[4934]: I1227 07:42:26.353213 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:26 crc kubenswrapper[4934]: I1227 07:42:26.354557 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:26 crc kubenswrapper[4934]: I1227 07:42:26.354621 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:26 crc kubenswrapper[4934]: I1227 07:42:26.354647 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:26 crc kubenswrapper[4934]: I1227 07:42:26.413287 4934 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 06:28:13.810206024 +0000 UTC Dec 27 07:42:26 crc kubenswrapper[4934]: I1227 07:42:26.413337 4934 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 478h45m47.396874094s for next certificate rotation Dec 27 07:42:26 crc kubenswrapper[4934]: I1227 07:42:26.532634 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:26 crc kubenswrapper[4934]: I1227 07:42:26.534253 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:26 crc kubenswrapper[4934]: I1227 07:42:26.534313 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:26 crc kubenswrapper[4934]: I1227 07:42:26.534337 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:27 crc kubenswrapper[4934]: I1227 07:42:27.143689 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:42:27 crc kubenswrapper[4934]: I1227 07:42:27.143929 4934 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 27 07:42:27 crc kubenswrapper[4934]: I1227 07:42:27.143995 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:27 crc kubenswrapper[4934]: I1227 07:42:27.145699 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:27 crc kubenswrapper[4934]: I1227 07:42:27.145766 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:27 crc kubenswrapper[4934]: I1227 07:42:27.145783 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:27 crc kubenswrapper[4934]: I1227 07:42:27.534465 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:27 crc kubenswrapper[4934]: I1227 07:42:27.535543 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:27 crc kubenswrapper[4934]: I1227 07:42:27.535601 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:27 crc kubenswrapper[4934]: I1227 07:42:27.535622 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:29 crc kubenswrapper[4934]: I1227 07:42:29.349949 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:42:29 crc kubenswrapper[4934]: I1227 07:42:29.350238 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:29 crc kubenswrapper[4934]: I1227 07:42:29.351877 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:29 crc kubenswrapper[4934]: I1227 07:42:29.351935 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:29 crc kubenswrapper[4934]: I1227 07:42:29.351955 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:29 crc kubenswrapper[4934]: E1227 07:42:29.566179 4934 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 27 07:42:31 crc kubenswrapper[4934]: E1227 07:42:31.069556 4934 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 27 07:42:31 crc kubenswrapper[4934]: I1227 07:42:31.079556 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:42:31 crc kubenswrapper[4934]: I1227 07:42:31.079786 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:31 crc kubenswrapper[4934]: I1227 07:42:31.081709 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:31 crc kubenswrapper[4934]: I1227 07:42:31.081769 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:31 crc kubenswrapper[4934]: I1227 07:42:31.081789 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:31 crc kubenswrapper[4934]: I1227 07:42:31.088468 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:42:31 crc kubenswrapper[4934]: I1227 07:42:31.406942 4934 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 27 07:42:31 crc kubenswrapper[4934]: E1227 07:42:31.443475 4934 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 27 07:42:31 crc kubenswrapper[4934]: I1227 07:42:31.547583 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:31 crc kubenswrapper[4934]: I1227 07:42:31.548682 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:31 crc kubenswrapper[4934]: I1227 07:42:31.548720 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:31 crc kubenswrapper[4934]: I1227 07:42:31.548730 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:32 crc kubenswrapper[4934]: I1227 07:42:32.233612 4934 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 27 07:42:32 crc kubenswrapper[4934]: I1227 07:42:32.233693 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 27 07:42:32 crc kubenswrapper[4934]: I1227 07:42:32.237350 4934 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 27 07:42:32 crc kubenswrapper[4934]: I1227 07:42:32.237409 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 27 07:42:32 crc kubenswrapper[4934]: I1227 07:42:32.578774 4934 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]log ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]etcd ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/generic-apiserver-start-informers ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/priority-and-fairness-filter ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/start-apiextensions-informers ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/start-apiextensions-controllers ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/crd-informer-synced ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/start-system-namespaces-controller ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 27 07:42:32 crc kubenswrapper[4934]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 27 07:42:32 crc kubenswrapper[4934]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/bootstrap-controller ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/start-kube-aggregator-informers ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/apiservice-registration-controller ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/apiservice-discovery-controller ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]autoregister-completion ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/apiservice-openapi-controller ok Dec 27 07:42:32 crc kubenswrapper[4934]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 27 07:42:32 crc kubenswrapper[4934]: livez check failed Dec 27 07:42:32 crc kubenswrapper[4934]: I1227 07:42:32.578861 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 07:42:32 crc kubenswrapper[4934]: I1227 07:42:32.670139 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:32 crc kubenswrapper[4934]: I1227 07:42:32.671510 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:32 crc kubenswrapper[4934]: I1227 07:42:32.671554 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:32 crc kubenswrapper[4934]: I1227 07:42:32.671565 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:32 crc kubenswrapper[4934]: I1227 07:42:32.671596 4934 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 27 07:42:35 crc kubenswrapper[4934]: I1227 07:42:35.596866 4934 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 27 07:42:35 crc kubenswrapper[4934]: I1227 07:42:35.614150 4934 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 27 07:42:35 crc kubenswrapper[4934]: I1227 07:42:35.685311 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 27 07:42:35 crc kubenswrapper[4934]: I1227 07:42:35.685592 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:35 crc kubenswrapper[4934]: I1227 07:42:35.687270 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:35 crc kubenswrapper[4934]: I1227 07:42:35.687353 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:35 crc kubenswrapper[4934]: I1227 07:42:35.687379 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:35 crc kubenswrapper[4934]: I1227 07:42:35.709166 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 27 07:42:36 crc kubenswrapper[4934]: I1227 07:42:36.160415 4934 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 07:42:36 crc kubenswrapper[4934]: I1227 07:42:36.160547 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 07:42:36 crc kubenswrapper[4934]: I1227 07:42:36.560344 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:36 crc kubenswrapper[4934]: I1227 07:42:36.564816 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:36 crc kubenswrapper[4934]: I1227 07:42:36.564876 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:36 crc kubenswrapper[4934]: I1227 07:42:36.564898 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:36 crc kubenswrapper[4934]: I1227 07:42:36.646131 4934 csr.go:261] certificate signing request csr-j6l46 is approved, waiting to be issued Dec 27 07:42:36 crc kubenswrapper[4934]: I1227 07:42:36.659530 4934 csr.go:257] certificate signing request csr-j6l46 is issued Dec 27 07:42:37 crc kubenswrapper[4934]: E1227 07:42:37.224749 4934 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.226735 4934 trace.go:236] Trace[1964469447]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Dec-2025 07:42:22.526) (total time: 14700ms): Dec 27 07:42:37 crc kubenswrapper[4934]: Trace[1964469447]: ---"Objects listed" error: 14700ms (07:42:37.226) Dec 27 07:42:37 crc kubenswrapper[4934]: Trace[1964469447]: [14.700072171s] [14.700072171s] END Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.226765 4934 trace.go:236] Trace[1274546249]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Dec-2025 07:42:22.783) (total time: 14443ms): Dec 27 07:42:37 crc kubenswrapper[4934]: Trace[1274546249]: ---"Objects listed" error: 14443ms (07:42:37.226) Dec 27 07:42:37 crc kubenswrapper[4934]: Trace[1274546249]: [14.443151453s] [14.443151453s] END Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.226779 4934 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.226795 4934 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.228678 4934 trace.go:236] Trace[1622362540]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Dec-2025 07:42:24.142) (total time: 13086ms): Dec 27 07:42:37 crc kubenswrapper[4934]: Trace[1622362540]: ---"Objects listed" error: 13086ms (07:42:37.228) Dec 27 07:42:37 crc kubenswrapper[4934]: Trace[1622362540]: [13.086122303s] [13.086122303s] END Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.228700 4934 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.229293 4934 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.229587 4934 trace.go:236] Trace[157690195]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Dec-2025 07:42:23.109) (total time: 14120ms): Dec 27 07:42:37 crc kubenswrapper[4934]: Trace[157690195]: ---"Objects listed" error: 14119ms (07:42:37.229) Dec 27 07:42:37 crc kubenswrapper[4934]: Trace[157690195]: [14.120083312s] [14.120083312s] END Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.229637 4934 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.415272 4934 apiserver.go:52] "Watching apiserver" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.418220 4934 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.418564 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.419075 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.419163 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.419201 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:37 crc kubenswrapper[4934]: E1227 07:42:37.419264 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.419330 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.419446 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:42:37 crc kubenswrapper[4934]: E1227 07:42:37.419514 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.419551 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:42:37 crc kubenswrapper[4934]: E1227 07:42:37.419600 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.420867 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.421350 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.422690 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.422712 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.422697 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.422745 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.422764 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.422936 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.423134 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.453071 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.468656 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.485315 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.498853 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.507286 4934 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.507738 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.515252 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.529398 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.530657 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.531027 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.531283 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.530971 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.531231 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.531376 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.531402 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532005 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532091 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532127 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532152 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532171 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532190 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532208 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532225 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532241 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532257 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532277 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532293 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532311 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532331 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532347 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532363 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532383 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532401 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532414 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532412 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532420 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532484 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532492 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532561 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532599 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532606 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532646 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532654 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532680 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532708 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532715 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532729 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532756 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532766 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532783 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532806 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532777 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532832 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532854 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532867 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532879 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532904 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532926 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532938 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.532972 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533031 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533049 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533057 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533084 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533126 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533151 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533174 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533195 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533218 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533219 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533242 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533257 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533264 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533287 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533311 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533322 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533332 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533385 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533394 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533386 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533429 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533449 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533503 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533528 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533544 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533552 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533579 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533607 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533615 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533624 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533648 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533687 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533711 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533731 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533744 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533767 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533802 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533836 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533869 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533905 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533936 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533937 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533947 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533958 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533975 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.533981 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534002 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534019 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534035 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534051 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534056 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534067 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534092 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534121 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534139 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534157 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534171 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534173 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534187 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534205 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534221 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534237 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534253 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534269 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534286 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534302 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534317 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534332 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534349 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534365 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534380 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534398 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534417 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534431 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534448 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534464 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534480 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534495 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534511 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534527 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534544 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534559 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534576 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534595 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534613 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534630 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534645 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534663 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534679 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534693 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534708 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534724 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534740 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534757 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534775 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534840 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534857 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534875 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534890 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534905 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534923 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534940 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534176 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534194 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534217 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534362 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534385 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534444 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534474 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534545 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534579 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534627 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534730 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534743 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534763 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534793 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534901 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: E1227 07:42:37.534967 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:42:38.034950005 +0000 UTC m=+18.855390599 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541142 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541193 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541220 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541246 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541272 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541298 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541324 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541346 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541371 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541395 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541417 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541443 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541466 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541488 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541514 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541535 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541558 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541581 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541633 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541654 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541678 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541700 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541721 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541743 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541766 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541790 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541819 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541850 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541885 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541919 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541954 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.541982 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542008 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542032 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542055 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542085 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542145 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542154 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542189 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542199 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542226 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542251 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542494 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542541 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542575 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542608 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542645 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542680 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542711 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542742 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542776 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542790 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542809 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542843 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542877 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542908 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542940 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542977 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.543010 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.542949 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.543043 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.543125 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.543151 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.543168 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.543382 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.543600 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.543645 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.543682 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.543716 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.543750 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.543791 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.543827 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.543864 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.543899 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.543974 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.544014 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.544050 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.545089 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.545154 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.545191 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.545232 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.545267 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.545303 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.545369 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.545414 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.545450 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.545485 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.545611 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.546342 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.546385 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.546421 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.546456 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.546486 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.546518 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.546553 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.546590 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.546738 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.543742 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.543977 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534977 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534995 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.535129 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.535162 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.535161 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.535189 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.535214 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.535321 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.535363 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.535389 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.535391 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.535398 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.535450 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.535501 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.535549 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.535599 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.535674 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.535668 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.535716 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.535834 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.535848 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.536153 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.536766 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.536939 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.537050 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.537076 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.537327 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.538625 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.538861 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.539157 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.539383 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.539693 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.539712 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.539850 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.539875 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.540274 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.544116 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.543331 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.544532 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.544832 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.544837 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.534967 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.545018 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.545069 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.545740 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.545987 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.546273 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.546228 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.546532 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.546598 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.546797 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: E1227 07:42:37.549686 4934 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 27 07:42:37 crc kubenswrapper[4934]: E1227 07:42:37.549791 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:38.049764875 +0000 UTC m=+18.870205489 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.550345 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.550909 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.551064 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.551213 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.551372 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.551608 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.551699 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.551953 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.552131 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.552177 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.552360 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.552366 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.552544 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.552568 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.552596 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.552748 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.552951 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.552989 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.553295 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.553498 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.553521 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: E1227 07:42:37.553538 4934 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.553736 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: E1227 07:42:37.554035 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:38.053862255 +0000 UTC m=+18.874302889 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.554265 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.554324 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.554505 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.554739 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.554832 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.555180 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.555282 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.555288 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.555340 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.555388 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.555438 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.555662 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.555678 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.555704 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.555820 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.555968 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556003 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556126 4934 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556149 4934 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556170 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556183 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556194 4934 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556206 4934 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556217 4934 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556229 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556245 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556289 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556304 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556320 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556331 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556344 4934 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556388 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556399 4934 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556411 4934 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556423 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556428 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556435 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556499 4934 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556513 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556524 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556534 4934 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556543 4934 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556553 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556562 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556571 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556582 4934 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556594 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556606 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556615 4934 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556624 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556633 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556644 4934 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556654 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556752 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556767 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556779 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.556788 4934 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558252 4934 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558271 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558281 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558292 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558304 4934 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558312 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558321 4934 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558330 4934 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558338 4934 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558347 4934 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558357 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558366 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558375 4934 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558384 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558392 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558400 4934 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558410 4934 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558419 4934 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558430 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558439 4934 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558447 4934 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558455 4934 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558463 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558471 4934 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558480 4934 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558488 4934 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558497 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558506 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558514 4934 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558523 4934 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558531 4934 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558540 4934 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558548 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558557 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558566 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558575 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558583 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558592 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558602 4934 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558611 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.557064 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558620 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558636 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558645 4934 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558654 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558664 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558672 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558682 4934 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558690 4934 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558698 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558707 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558716 4934 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558725 4934 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558734 4934 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558779 4934 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558809 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558817 4934 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558825 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558833 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.557527 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558155 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.558604 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559109 4934 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.557190 4934 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559194 4934 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559393 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559403 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559413 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559424 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559432 4934 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559442 4934 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559450 4934 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559458 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559466 4934 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559476 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559485 4934 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559494 4934 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559502 4934 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559511 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559520 4934 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559528 4934 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559537 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559545 4934 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559553 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559562 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559570 4934 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559579 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559590 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559599 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559607 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559616 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559624 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559633 4934 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559642 4934 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559666 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559676 4934 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559686 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559694 4934 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559702 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559710 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559718 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559727 4934 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559735 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559744 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559752 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559760 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559769 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559777 4934 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559785 4934 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559793 4934 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.559861 4934 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.565293 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.565321 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.565679 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.566168 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.568471 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.572527 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.573766 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.574671 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: E1227 07:42:37.574890 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 27 07:42:37 crc kubenswrapper[4934]: E1227 07:42:37.574968 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 27 07:42:37 crc kubenswrapper[4934]: E1227 07:42:37.574997 4934 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:37 crc kubenswrapper[4934]: E1227 07:42:37.575134 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:38.075067528 +0000 UTC m=+18.895508222 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:37 crc kubenswrapper[4934]: E1227 07:42:37.575659 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 27 07:42:37 crc kubenswrapper[4934]: E1227 07:42:37.575689 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 27 07:42:37 crc kubenswrapper[4934]: E1227 07:42:37.575701 4934 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:37 crc kubenswrapper[4934]: E1227 07:42:37.575756 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:38.075735946 +0000 UTC m=+18.896176530 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.575778 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.576087 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.576289 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.576898 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.577905 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.578275 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.578949 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.579058 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.579244 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.580519 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.584155 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.584909 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.585255 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.585456 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.586298 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.586325 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.586436 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.586454 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.587315 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.587402 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.587532 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.587546 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.587731 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.588435 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.590029 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.592734 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.594047 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.595119 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.595258 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.595482 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.596127 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.601614 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.601760 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.601983 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.602092 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.602185 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.602460 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.603888 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.604305 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.604338 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.604494 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.604643 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.605067 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.605185 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.605860 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.609481 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.613355 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.617361 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.626728 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.634346 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.641378 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.651933 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.660728 4934 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-27 07:37:36 +0000 UTC, rotation deadline is 2026-10-16 00:29:25.891575885 +0000 UTC Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.660813 4934 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7024h46m48.230768163s for next certificate rotation Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.661343 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.661494 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.661722 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.661891 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.661990 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.662106 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.662357 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.662457 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.662548 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.662628 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.662304 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.662010 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.662864 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.662920 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.662971 4934 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.663020 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.663068 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.663157 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.663219 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.663274 4934 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.663329 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.663378 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.663440 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.663500 4934 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.663552 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.663604 4934 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.663661 4934 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.663738 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.663791 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.663848 4934 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.663901 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.663951 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.663999 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.664048 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.664117 4934 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.664213 4934 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.664271 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.664800 4934 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.664872 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.664925 4934 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.664984 4934 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.665042 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.665112 4934 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.665183 4934 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.665256 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.665320 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.665441 4934 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.665496 4934 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.665549 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.665604 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.665660 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.665713 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.665763 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.665812 4934 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.672985 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.673475 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: E1227 07:42:37.681452 4934 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.690656 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.719033 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.739277 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.741382 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.746150 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.754377 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 27 07:42:37 crc kubenswrapper[4934]: W1227 07:42:37.759050 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-4a33b51c3e472f767de1ccf91f34202bbaf870137f5b6941facaef667e19e33e WatchSource:0}: Error finding container 4a33b51c3e472f767de1ccf91f34202bbaf870137f5b6941facaef667e19e33e: Status 404 returned error can't find the container with id 4a33b51c3e472f767de1ccf91f34202bbaf870137f5b6941facaef667e19e33e Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.763720 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.780001 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.801135 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:37 crc kubenswrapper[4934]: I1227 07:42:37.816410 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.070121 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.070218 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.070259 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:38 crc kubenswrapper[4934]: E1227 07:42:38.070323 4934 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 27 07:42:38 crc kubenswrapper[4934]: E1227 07:42:38.070350 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:42:39.070311576 +0000 UTC m=+19.890752170 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:42:38 crc kubenswrapper[4934]: E1227 07:42:38.070388 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:39.070378538 +0000 UTC m=+19.890819272 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 27 07:42:38 crc kubenswrapper[4934]: E1227 07:42:38.070413 4934 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 27 07:42:38 crc kubenswrapper[4934]: E1227 07:42:38.070514 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:39.070492351 +0000 UTC m=+19.890932945 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.171119 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.171195 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:42:38 crc kubenswrapper[4934]: E1227 07:42:38.171313 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 27 07:42:38 crc kubenswrapper[4934]: E1227 07:42:38.171352 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 27 07:42:38 crc kubenswrapper[4934]: E1227 07:42:38.171364 4934 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:38 crc kubenswrapper[4934]: E1227 07:42:38.171417 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:39.171401004 +0000 UTC m=+19.991841598 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:38 crc kubenswrapper[4934]: E1227 07:42:38.171323 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 27 07:42:38 crc kubenswrapper[4934]: E1227 07:42:38.171456 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 27 07:42:38 crc kubenswrapper[4934]: E1227 07:42:38.171470 4934 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:38 crc kubenswrapper[4934]: E1227 07:42:38.171531 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:39.171514107 +0000 UTC m=+19.991954711 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.415946 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-fwcxz"] Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.416222 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-fwcxz" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.418699 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.420913 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.424803 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.446302 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.466819 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.473610 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/72743c37-fd5a-4aa8-959b-6a1fd048c650-hosts-file\") pod \"node-resolver-fwcxz\" (UID: \"72743c37-fd5a-4aa8-959b-6a1fd048c650\") " pod="openshift-dns/node-resolver-fwcxz" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.473650 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxfqx\" (UniqueName: \"kubernetes.io/projected/72743c37-fd5a-4aa8-959b-6a1fd048c650-kube-api-access-rxfqx\") pod \"node-resolver-fwcxz\" (UID: \"72743c37-fd5a-4aa8-959b-6a1fd048c650\") " pod="openshift-dns/node-resolver-fwcxz" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.489775 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.512881 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.525735 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.545554 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.556553 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.568552 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.570976 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"85a48c1006a3175f5372e678a3ee480e8957f68bbeff2ded8447837605b08881"} Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.572356 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd"} Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.572413 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e"} Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.572425 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4a33b51c3e472f767de1ccf91f34202bbaf870137f5b6941facaef667e19e33e"} Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.573519 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8"} Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.573552 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"804de5a1f3e8f1eccd7e6351f9ae0556f9821ec8808020f882957ef6ae2f93dc"} Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.573899 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/72743c37-fd5a-4aa8-959b-6a1fd048c650-hosts-file\") pod \"node-resolver-fwcxz\" (UID: \"72743c37-fd5a-4aa8-959b-6a1fd048c650\") " pod="openshift-dns/node-resolver-fwcxz" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.573928 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxfqx\" (UniqueName: \"kubernetes.io/projected/72743c37-fd5a-4aa8-959b-6a1fd048c650-kube-api-access-rxfqx\") pod \"node-resolver-fwcxz\" (UID: \"72743c37-fd5a-4aa8-959b-6a1fd048c650\") " pod="openshift-dns/node-resolver-fwcxz" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.574006 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/72743c37-fd5a-4aa8-959b-6a1fd048c650-hosts-file\") pod \"node-resolver-fwcxz\" (UID: \"72743c37-fd5a-4aa8-959b-6a1fd048c650\") " pod="openshift-dns/node-resolver-fwcxz" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.586438 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.588674 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxfqx\" (UniqueName: \"kubernetes.io/projected/72743c37-fd5a-4aa8-959b-6a1fd048c650-kube-api-access-rxfqx\") pod \"node-resolver-fwcxz\" (UID: \"72743c37-fd5a-4aa8-959b-6a1fd048c650\") " pod="openshift-dns/node-resolver-fwcxz" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.601045 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.623474 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.647941 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.678159 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.690807 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.699165 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.711407 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.725183 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.727317 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-fwcxz" Dec 27 07:42:38 crc kubenswrapper[4934]: W1227 07:42:38.738700 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72743c37_fd5a_4aa8_959b_6a1fd048c650.slice/crio-0e45982e2129c748df286124bba8ba108f55ca70e8d5dc6f563cb998c1830884 WatchSource:0}: Error finding container 0e45982e2129c748df286124bba8ba108f55ca70e8d5dc6f563cb998c1830884: Status 404 returned error can't find the container with id 0e45982e2129c748df286124bba8ba108f55ca70e8d5dc6f563cb998c1830884 Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.744246 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.755855 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.770797 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.785606 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.796963 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.808891 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:38 crc kubenswrapper[4934]: I1227 07:42:38.819326 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.077895 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.077995 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.078023 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:39 crc kubenswrapper[4934]: E1227 07:42:39.078099 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:42:41.078068268 +0000 UTC m=+21.898508862 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:42:39 crc kubenswrapper[4934]: E1227 07:42:39.078160 4934 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 27 07:42:39 crc kubenswrapper[4934]: E1227 07:42:39.078186 4934 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 27 07:42:39 crc kubenswrapper[4934]: E1227 07:42:39.078242 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:41.078230182 +0000 UTC m=+21.898670776 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 27 07:42:39 crc kubenswrapper[4934]: E1227 07:42:39.078287 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:41.078276903 +0000 UTC m=+21.898717497 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.179133 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.179209 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:42:39 crc kubenswrapper[4934]: E1227 07:42:39.179337 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 27 07:42:39 crc kubenswrapper[4934]: E1227 07:42:39.179357 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 27 07:42:39 crc kubenswrapper[4934]: E1227 07:42:39.179371 4934 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:39 crc kubenswrapper[4934]: E1227 07:42:39.179425 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:41.179407953 +0000 UTC m=+21.999848547 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:39 crc kubenswrapper[4934]: E1227 07:42:39.179804 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 27 07:42:39 crc kubenswrapper[4934]: E1227 07:42:39.179828 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 27 07:42:39 crc kubenswrapper[4934]: E1227 07:42:39.179838 4934 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:39 crc kubenswrapper[4934]: E1227 07:42:39.179868 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:41.179859305 +0000 UTC m=+22.000299899 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.188709 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-w9j4r"] Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.189126 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.191750 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.192108 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.192115 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.192120 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.192502 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.208123 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.219760 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.233029 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.242776 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.252658 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.269463 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.279184 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.279580 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7-mcd-auth-proxy-config\") pod \"machine-config-daemon-w9j4r\" (UID: \"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\") " pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.279622 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c572j\" (UniqueName: \"kubernetes.io/projected/a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7-kube-api-access-c572j\") pod \"machine-config-daemon-w9j4r\" (UID: \"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\") " pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.279655 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7-proxy-tls\") pod \"machine-config-daemon-w9j4r\" (UID: \"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\") " pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.279689 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7-rootfs\") pod \"machine-config-daemon-w9j4r\" (UID: \"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\") " pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.287415 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.296344 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.329530 4934 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 27 07:42:39 crc kubenswrapper[4934]: W1227 07:42:39.329749 4934 reflector.go:484] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 27 07:42:39 crc kubenswrapper[4934]: W1227 07:42:39.329812 4934 reflector.go:484] object-"openshift-dns"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 27 07:42:39 crc kubenswrapper[4934]: W1227 07:42:39.330115 4934 reflector.go:484] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": watch of *v1.Secret ended with: very short watch: object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": Unexpected watch close - watch lasted less than a second and no items received Dec 27 07:42:39 crc kubenswrapper[4934]: W1227 07:42:39.330162 4934 reflector.go:484] object-"openshift-machine-config-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 27 07:42:39 crc kubenswrapper[4934]: W1227 07:42:39.330195 4934 reflector.go:484] object-"openshift-machine-config-operator"/"proxy-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"proxy-tls": Unexpected watch close - watch lasted less than a second and no items received Dec 27 07:42:39 crc kubenswrapper[4934]: W1227 07:42:39.330362 4934 reflector.go:484] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": Unexpected watch close - watch lasted less than a second and no items received Dec 27 07:42:39 crc kubenswrapper[4934]: W1227 07:42:39.330393 4934 reflector.go:484] object-"openshift-machine-config-operator"/"kube-rbac-proxy": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-rbac-proxy": Unexpected watch close - watch lasted less than a second and no items received Dec 27 07:42:39 crc kubenswrapper[4934]: W1227 07:42:39.330413 4934 reflector.go:484] object-"openshift-dns"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.380221 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7-rootfs\") pod \"machine-config-daemon-w9j4r\" (UID: \"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\") " pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.380279 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7-mcd-auth-proxy-config\") pod \"machine-config-daemon-w9j4r\" (UID: \"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\") " pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.380297 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c572j\" (UniqueName: \"kubernetes.io/projected/a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7-kube-api-access-c572j\") pod \"machine-config-daemon-w9j4r\" (UID: \"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\") " pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.380312 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7-proxy-tls\") pod \"machine-config-daemon-w9j4r\" (UID: \"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\") " pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.380372 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7-rootfs\") pod \"machine-config-daemon-w9j4r\" (UID: \"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\") " pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.381124 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7-mcd-auth-proxy-config\") pod \"machine-config-daemon-w9j4r\" (UID: \"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\") " pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.386245 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7-proxy-tls\") pod \"machine-config-daemon-w9j4r\" (UID: \"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\") " pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.394418 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c572j\" (UniqueName: \"kubernetes.io/projected/a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7-kube-api-access-c572j\") pod \"machine-config-daemon-w9j4r\" (UID: \"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\") " pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.469482 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.469589 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:42:39 crc kubenswrapper[4934]: E1227 07:42:39.469632 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.469688 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:39 crc kubenswrapper[4934]: E1227 07:42:39.469842 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:42:39 crc kubenswrapper[4934]: E1227 07:42:39.469928 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.477020 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.478386 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.481500 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.483325 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.485718 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.485887 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.487468 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.489190 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.491424 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.492883 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.494315 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.495451 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.496836 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.497892 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.499012 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.500162 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.500290 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.500318 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.502496 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.504131 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.506271 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.507620 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.509690 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.510390 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.511200 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.512474 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.513404 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.514468 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.514950 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.515056 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.516182 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.516622 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.517585 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.518059 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.518502 4934 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.518596 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.520507 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.520955 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.521786 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.523317 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.524049 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.525130 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.525881 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.526028 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.527233 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.527964 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.529202 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.529919 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.531037 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.531607 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.532640 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.533437 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.534707 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.535729 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.536531 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.537563 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.538310 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.539016 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.540065 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.542492 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.556559 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.564356 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-jzv2s"] Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.564757 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.565964 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8vzdl"] Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.572939 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.573261 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.573360 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.574048 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.580670 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.581085 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-hqfw4"] Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.581478 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.581910 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.582461 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.585587 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.585713 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.586192 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.586322 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.586208 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.586393 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.586578 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.586676 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.586767 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"2acb16dad1fe06534ff430b29f3c95d423811762b23cf4a1ae07c35235d01cd4"} Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.586815 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.589188 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-fwcxz" event={"ID":"72743c37-fd5a-4aa8-959b-6a1fd048c650","Type":"ContainerStarted","Data":"d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee"} Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.589318 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-fwcxz" event={"ID":"72743c37-fd5a-4aa8-959b-6a1fd048c650","Type":"ContainerStarted","Data":"0e45982e2129c748df286124bba8ba108f55ca70e8d5dc6f563cb998c1830884"} Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.597286 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.614577 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.630836 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.647385 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.657246 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.669367 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.681514 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.683541 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-host-run-multus-certs\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.683576 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-run-ovn-kubernetes\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.683593 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1bdd4b78-f615-4771-9448-8faa03dbc4f5-cni-binary-copy\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.683610 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-multus-socket-dir-parent\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.683626 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-host-run-k8s-cni-cncf-io\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.683654 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1c4780a7-56b7-4772-af66-045a03b31a71-ovnkube-script-lib\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.683700 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kp9j\" (UniqueName: \"kubernetes.io/projected/1c4780a7-56b7-4772-af66-045a03b31a71-kube-api-access-2kp9j\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.683723 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-etc-openvswitch\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.683740 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-node-log\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.683796 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e59598db-f0ef-42fb-b151-b436ac250b08-tuning-conf-dir\") pod \"multus-additional-cni-plugins-hqfw4\" (UID: \"e59598db-f0ef-42fb-b151-b436ac250b08\") " pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.683812 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4ktc\" (UniqueName: \"kubernetes.io/projected/e59598db-f0ef-42fb-b151-b436ac250b08-kube-api-access-r4ktc\") pod \"multus-additional-cni-plugins-hqfw4\" (UID: \"e59598db-f0ef-42fb-b151-b436ac250b08\") " pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.683829 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-run-ovn\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.683860 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-multus-cni-dir\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.683891 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-host-var-lib-cni-bin\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.683906 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-host-var-lib-kubelet\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.683923 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-kubelet\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.683942 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-systemd-units\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.683962 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e59598db-f0ef-42fb-b151-b436ac250b08-system-cni-dir\") pod \"multus-additional-cni-plugins-hqfw4\" (UID: \"e59598db-f0ef-42fb-b151-b436ac250b08\") " pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.683979 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-run-netns\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.683993 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-cni-netd\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684009 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1c4780a7-56b7-4772-af66-045a03b31a71-env-overrides\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684023 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1c4780a7-56b7-4772-af66-045a03b31a71-ovn-node-metrics-cert\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684037 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-system-cni-dir\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684052 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-host-run-netns\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684067 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-hostroot\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684101 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1bdd4b78-f615-4771-9448-8faa03dbc4f5-multus-daemon-config\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684124 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s295j\" (UniqueName: \"kubernetes.io/projected/1bdd4b78-f615-4771-9448-8faa03dbc4f5-kube-api-access-s295j\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684139 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e59598db-f0ef-42fb-b151-b436ac250b08-os-release\") pod \"multus-additional-cni-plugins-hqfw4\" (UID: \"e59598db-f0ef-42fb-b151-b436ac250b08\") " pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684154 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-log-socket\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684170 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e59598db-f0ef-42fb-b151-b436ac250b08-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-hqfw4\" (UID: \"e59598db-f0ef-42fb-b151-b436ac250b08\") " pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684184 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-cni-bin\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684199 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1c4780a7-56b7-4772-af66-045a03b31a71-ovnkube-config\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684213 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e59598db-f0ef-42fb-b151-b436ac250b08-cni-binary-copy\") pod \"multus-additional-cni-plugins-hqfw4\" (UID: \"e59598db-f0ef-42fb-b151-b436ac250b08\") " pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684229 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-slash\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684259 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-host-var-lib-cni-multus\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684274 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-multus-conf-dir\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684291 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e59598db-f0ef-42fb-b151-b436ac250b08-cnibin\") pod \"multus-additional-cni-plugins-hqfw4\" (UID: \"e59598db-f0ef-42fb-b151-b436ac250b08\") " pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684304 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-run-systemd\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684318 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-cnibin\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684332 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-var-lib-openvswitch\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684346 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-run-openvswitch\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684362 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684377 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-os-release\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.684391 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-etc-kubernetes\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.696774 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.709301 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.722567 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.733796 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.751043 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.761691 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.773786 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.785425 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e59598db-f0ef-42fb-b151-b436ac250b08-tuning-conf-dir\") pod \"multus-additional-cni-plugins-hqfw4\" (UID: \"e59598db-f0ef-42fb-b151-b436ac250b08\") " pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.785479 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4ktc\" (UniqueName: \"kubernetes.io/projected/e59598db-f0ef-42fb-b151-b436ac250b08-kube-api-access-r4ktc\") pod \"multus-additional-cni-plugins-hqfw4\" (UID: \"e59598db-f0ef-42fb-b151-b436ac250b08\") " pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.785517 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-run-ovn\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.785580 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-multus-cni-dir\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.785619 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-host-var-lib-cni-bin\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.785664 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-host-var-lib-kubelet\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.785704 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-kubelet\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.785725 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-run-ovn\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.785750 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-systemd-units\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.785787 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-cni-netd\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.785810 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-kubelet\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.785819 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1c4780a7-56b7-4772-af66-045a03b31a71-env-overrides\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.785844 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-systemd-units\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.785838 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-host-var-lib-kubelet\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.785881 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-cni-netd\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.785852 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1c4780a7-56b7-4772-af66-045a03b31a71-ovn-node-metrics-cert\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.785938 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-multus-cni-dir\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.785838 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-host-var-lib-cni-bin\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.785952 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e59598db-f0ef-42fb-b151-b436ac250b08-system-cni-dir\") pod \"multus-additional-cni-plugins-hqfw4\" (UID: \"e59598db-f0ef-42fb-b151-b436ac250b08\") " pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786026 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-run-netns\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.785986 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e59598db-f0ef-42fb-b151-b436ac250b08-system-cni-dir\") pod \"multus-additional-cni-plugins-hqfw4\" (UID: \"e59598db-f0ef-42fb-b151-b436ac250b08\") " pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786065 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-host-run-netns\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786141 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-host-run-netns\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786141 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-run-netns\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786176 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-system-cni-dir\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786221 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s295j\" (UniqueName: \"kubernetes.io/projected/1bdd4b78-f615-4771-9448-8faa03dbc4f5-kube-api-access-s295j\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786253 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e59598db-f0ef-42fb-b151-b436ac250b08-os-release\") pod \"multus-additional-cni-plugins-hqfw4\" (UID: \"e59598db-f0ef-42fb-b151-b436ac250b08\") " pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786283 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-log-socket\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786327 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-system-cni-dir\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786349 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-hostroot\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786380 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1bdd4b78-f615-4771-9448-8faa03dbc4f5-multus-daemon-config\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786428 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e59598db-f0ef-42fb-b151-b436ac250b08-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-hqfw4\" (UID: \"e59598db-f0ef-42fb-b151-b436ac250b08\") " pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786462 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-cni-bin\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786492 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1c4780a7-56b7-4772-af66-045a03b31a71-ovnkube-config\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786525 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e59598db-f0ef-42fb-b151-b436ac250b08-cni-binary-copy\") pod \"multus-additional-cni-plugins-hqfw4\" (UID: \"e59598db-f0ef-42fb-b151-b436ac250b08\") " pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786555 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-slash\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786584 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-host-var-lib-cni-multus\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786616 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-multus-conf-dir\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786650 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e59598db-f0ef-42fb-b151-b436ac250b08-cnibin\") pod \"multus-additional-cni-plugins-hqfw4\" (UID: \"e59598db-f0ef-42fb-b151-b436ac250b08\") " pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786663 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-hostroot\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786680 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-run-systemd\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786712 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-log-socket\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786730 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786766 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-cnibin\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786784 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e59598db-f0ef-42fb-b151-b436ac250b08-os-release\") pod \"multus-additional-cni-plugins-hqfw4\" (UID: \"e59598db-f0ef-42fb-b151-b436ac250b08\") " pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786800 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-var-lib-openvswitch\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786818 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-host-var-lib-cni-multus\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786833 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-run-openvswitch\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786880 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-run-openvswitch\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786879 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-os-release\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786930 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-etc-kubernetes\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786938 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-os-release\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.786936 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1c4780a7-56b7-4772-af66-045a03b31a71-env-overrides\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787016 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-host-run-multus-certs\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787048 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-slash\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787053 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-run-ovn-kubernetes\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787111 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-run-ovn-kubernetes\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787116 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-multus-socket-dir-parent\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787150 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-host-run-k8s-cni-cncf-io\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787182 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-cnibin\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787219 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-run-systemd\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787251 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-var-lib-openvswitch\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787218 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e59598db-f0ef-42fb-b151-b436ac250b08-cnibin\") pod \"multus-additional-cni-plugins-hqfw4\" (UID: \"e59598db-f0ef-42fb-b151-b436ac250b08\") " pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787214 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-cni-bin\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787303 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-host-run-k8s-cni-cncf-io\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787308 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-multus-socket-dir-parent\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787342 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-etc-kubernetes\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787184 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-multus-conf-dir\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787181 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1c4780a7-56b7-4772-af66-045a03b31a71-ovnkube-script-lib\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787421 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kp9j\" (UniqueName: \"kubernetes.io/projected/1c4780a7-56b7-4772-af66-045a03b31a71-kube-api-access-2kp9j\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787459 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1bdd4b78-f615-4771-9448-8faa03dbc4f5-cni-binary-copy\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787490 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-etc-openvswitch\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787518 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-node-log\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787605 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-node-log\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787613 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1bdd4b78-f615-4771-9448-8faa03dbc4f5-multus-daemon-config\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787151 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787722 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-etc-openvswitch\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787346 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1bdd4b78-f615-4771-9448-8faa03dbc4f5-host-run-multus-certs\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.787820 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e59598db-f0ef-42fb-b151-b436ac250b08-cni-binary-copy\") pod \"multus-additional-cni-plugins-hqfw4\" (UID: \"e59598db-f0ef-42fb-b151-b436ac250b08\") " pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.788327 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1c4780a7-56b7-4772-af66-045a03b31a71-ovnkube-config\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.788382 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1c4780a7-56b7-4772-af66-045a03b31a71-ovnkube-script-lib\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.788642 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e59598db-f0ef-42fb-b151-b436ac250b08-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-hqfw4\" (UID: \"e59598db-f0ef-42fb-b151-b436ac250b08\") " pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.788810 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1bdd4b78-f615-4771-9448-8faa03dbc4f5-cni-binary-copy\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.796956 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1c4780a7-56b7-4772-af66-045a03b31a71-ovn-node-metrics-cert\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.803742 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4ktc\" (UniqueName: \"kubernetes.io/projected/e59598db-f0ef-42fb-b151-b436ac250b08-kube-api-access-r4ktc\") pod \"multus-additional-cni-plugins-hqfw4\" (UID: \"e59598db-f0ef-42fb-b151-b436ac250b08\") " pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.803969 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s295j\" (UniqueName: \"kubernetes.io/projected/1bdd4b78-f615-4771-9448-8faa03dbc4f5-kube-api-access-s295j\") pod \"multus-jzv2s\" (UID: \"1bdd4b78-f615-4771-9448-8faa03dbc4f5\") " pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.817890 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kp9j\" (UniqueName: \"kubernetes.io/projected/1c4780a7-56b7-4772-af66-045a03b31a71-kube-api-access-2kp9j\") pod \"ovnkube-node-8vzdl\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.916428 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-jzv2s" Dec 27 07:42:39 crc kubenswrapper[4934]: I1227 07:42:39.931826 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:39 crc kubenswrapper[4934]: W1227 07:42:39.933448 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1bdd4b78_f615_4771_9448_8faa03dbc4f5.slice/crio-25c97b42fe4e3be3134177e82f1479d23d4fbec418a360aad65fea524d91c7be WatchSource:0}: Error finding container 25c97b42fe4e3be3134177e82f1479d23d4fbec418a360aad65fea524d91c7be: Status 404 returned error can't find the container with id 25c97b42fe4e3be3134177e82f1479d23d4fbec418a360aad65fea524d91c7be Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.004174 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e59598db-f0ef-42fb-b151-b436ac250b08-tuning-conf-dir\") pod \"multus-additional-cni-plugins-hqfw4\" (UID: \"e59598db-f0ef-42fb-b151-b436ac250b08\") " pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:40 crc kubenswrapper[4934]: W1227 07:42:40.017358 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c4780a7_56b7_4772_af66_045a03b31a71.slice/crio-18f078934d8adfcf20f5b70e00c58431e888b0eee3333dcf5da17bcf694a4684 WatchSource:0}: Error finding container 18f078934d8adfcf20f5b70e00c58431e888b0eee3333dcf5da17bcf694a4684: Status 404 returned error can't find the container with id 18f078934d8adfcf20f5b70e00c58431e888b0eee3333dcf5da17bcf694a4684 Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.225509 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" Dec 27 07:42:40 crc kubenswrapper[4934]: W1227 07:42:40.236742 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode59598db_f0ef_42fb_b151_b436ac250b08.slice/crio-d9e90aa139814a7004b5131f1bff2cc3c4161c4aa1bcab44c395425cd4bcd47e WatchSource:0}: Error finding container d9e90aa139814a7004b5131f1bff2cc3c4161c4aa1bcab44c395425cd4bcd47e: Status 404 returned error can't find the container with id d9e90aa139814a7004b5131f1bff2cc3c4161c4aa1bcab44c395425cd4bcd47e Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.501785 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.530303 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.533736 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.546949 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.560536 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.592395 4934 generic.go:334] "Generic (PLEG): container finished" podID="1c4780a7-56b7-4772-af66-045a03b31a71" containerID="92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd" exitCode=0 Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.592530 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerDied","Data":"92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd"} Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.592855 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerStarted","Data":"18f078934d8adfcf20f5b70e00c58431e888b0eee3333dcf5da17bcf694a4684"} Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.594945 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc"} Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.596552 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jzv2s" event={"ID":"1bdd4b78-f615-4771-9448-8faa03dbc4f5","Type":"ContainerStarted","Data":"a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44"} Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.596591 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jzv2s" event={"ID":"1bdd4b78-f615-4771-9448-8faa03dbc4f5","Type":"ContainerStarted","Data":"25c97b42fe4e3be3134177e82f1479d23d4fbec418a360aad65fea524d91c7be"} Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.598336 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae"} Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.598437 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c"} Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.599727 4934 generic.go:334] "Generic (PLEG): container finished" podID="e59598db-f0ef-42fb-b151-b436ac250b08" containerID="5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692" exitCode=0 Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.599813 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" event={"ID":"e59598db-f0ef-42fb-b151-b436ac250b08","Type":"ContainerDied","Data":"5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692"} Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.599848 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" event={"ID":"e59598db-f0ef-42fb-b151-b436ac250b08","Type":"ContainerStarted","Data":"d9e90aa139814a7004b5131f1bff2cc3c4161c4aa1bcab44c395425cd4bcd47e"} Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.607958 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.618472 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.635135 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.647056 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.655499 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.669436 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.670162 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.685361 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.697562 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.708917 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.720499 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.742708 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.745306 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.755997 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.767670 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.781797 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.784228 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.806069 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.819664 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.831923 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.845255 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.859399 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.872296 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.881765 4934 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.884390 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.884451 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.884471 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.884669 4934 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.886271 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.894439 4934 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.894803 4934 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.896300 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.896351 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.896365 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.896388 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.896404 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:40Z","lastTransitionTime":"2025-12-27T07:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.902998 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: E1227 07:42:40.910820 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.914411 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.914493 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.914510 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.914536 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.914559 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:40Z","lastTransitionTime":"2025-12-27T07:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.919455 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: E1227 07:42:40.925423 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.929224 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.929388 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.929505 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.929613 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.929697 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:40Z","lastTransitionTime":"2025-12-27T07:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.938387 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: E1227 07:42:40.942260 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.945402 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.945462 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.945473 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.945497 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.945509 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:40Z","lastTransitionTime":"2025-12-27T07:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:40 crc kubenswrapper[4934]: E1227 07:42:40.958709 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.962706 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.962755 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.962768 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.962789 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.962802 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:40Z","lastTransitionTime":"2025-12-27T07:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:40 crc kubenswrapper[4934]: E1227 07:42:40.973946 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:40Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:40 crc kubenswrapper[4934]: E1227 07:42:40.974172 4934 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.975874 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.975920 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.975931 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.975948 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:40 crc kubenswrapper[4934]: I1227 07:42:40.975963 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:40Z","lastTransitionTime":"2025-12-27T07:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.078896 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.078936 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.078948 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.078966 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.078979 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:41Z","lastTransitionTime":"2025-12-27T07:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.107576 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.107681 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.107731 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:41 crc kubenswrapper[4934]: E1227 07:42:41.107794 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:42:45.107758426 +0000 UTC m=+25.928199070 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:42:41 crc kubenswrapper[4934]: E1227 07:42:41.107843 4934 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 27 07:42:41 crc kubenswrapper[4934]: E1227 07:42:41.107866 4934 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 27 07:42:41 crc kubenswrapper[4934]: E1227 07:42:41.107904 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:45.107893209 +0000 UTC m=+25.928333863 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 27 07:42:41 crc kubenswrapper[4934]: E1227 07:42:41.107932 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:45.10791343 +0000 UTC m=+25.928354104 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.180756 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.181154 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.181166 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.181186 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.181198 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:41Z","lastTransitionTime":"2025-12-27T07:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.209123 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.209193 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:42:41 crc kubenswrapper[4934]: E1227 07:42:41.209343 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 27 07:42:41 crc kubenswrapper[4934]: E1227 07:42:41.209353 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 27 07:42:41 crc kubenswrapper[4934]: E1227 07:42:41.209388 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 27 07:42:41 crc kubenswrapper[4934]: E1227 07:42:41.209403 4934 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:41 crc kubenswrapper[4934]: E1227 07:42:41.209368 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 27 07:42:41 crc kubenswrapper[4934]: E1227 07:42:41.209477 4934 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:41 crc kubenswrapper[4934]: E1227 07:42:41.209463 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:45.20944308 +0000 UTC m=+26.029883774 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:41 crc kubenswrapper[4934]: E1227 07:42:41.209541 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:45.209524283 +0000 UTC m=+26.029964887 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.284006 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.284047 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.284059 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.284078 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.284120 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:41Z","lastTransitionTime":"2025-12-27T07:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.387607 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.387655 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.387668 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.387686 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.387698 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:41Z","lastTransitionTime":"2025-12-27T07:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.467150 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:42:41 crc kubenswrapper[4934]: E1227 07:42:41.467268 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.467571 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:42:41 crc kubenswrapper[4934]: E1227 07:42:41.467866 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.467918 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:41 crc kubenswrapper[4934]: E1227 07:42:41.467967 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.489545 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.489591 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.489604 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.489623 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.489634 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:41Z","lastTransitionTime":"2025-12-27T07:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.592862 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.593607 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.593622 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.593638 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.593647 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:41Z","lastTransitionTime":"2025-12-27T07:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.605359 4934 generic.go:334] "Generic (PLEG): container finished" podID="e59598db-f0ef-42fb-b151-b436ac250b08" containerID="170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0" exitCode=0 Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.605469 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" event={"ID":"e59598db-f0ef-42fb-b151-b436ac250b08","Type":"ContainerDied","Data":"170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0"} Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.613511 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerStarted","Data":"9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18"} Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.613559 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerStarted","Data":"d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c"} Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.613582 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerStarted","Data":"6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168"} Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.613598 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerStarted","Data":"b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6"} Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.624685 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.640453 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.656327 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.672762 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.682722 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.696437 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.696487 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.696502 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.696523 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.696536 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:41Z","lastTransitionTime":"2025-12-27T07:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.700003 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.708895 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.721441 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.747507 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.777542 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.793119 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.799410 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.799443 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.799451 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.799463 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.799472 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:41Z","lastTransitionTime":"2025-12-27T07:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.803206 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-dfj4x"] Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.803506 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-dfj4x" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.805590 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.805755 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.806118 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.806349 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.807062 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.818487 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.830004 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.838787 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.850145 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.862806 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.882085 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.895693 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.901016 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.901168 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.901233 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.901303 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.901359 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:41Z","lastTransitionTime":"2025-12-27T07:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.907084 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.915318 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f11f562a-9201-4bd4-9f28-a9b2d5cd44bd-serviceca\") pod \"node-ca-dfj4x\" (UID: \"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\") " pod="openshift-image-registry/node-ca-dfj4x" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.915357 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f11f562a-9201-4bd4-9f28-a9b2d5cd44bd-host\") pod \"node-ca-dfj4x\" (UID: \"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\") " pod="openshift-image-registry/node-ca-dfj4x" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.915374 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6bzl\" (UniqueName: \"kubernetes.io/projected/f11f562a-9201-4bd4-9f28-a9b2d5cd44bd-kube-api-access-k6bzl\") pod \"node-ca-dfj4x\" (UID: \"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\") " pod="openshift-image-registry/node-ca-dfj4x" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.919051 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.929326 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.940029 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.947742 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:41 crc kubenswrapper[4934]: I1227 07:42:41.958429 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:41Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.004174 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.004383 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.004445 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.004528 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.004585 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:42Z","lastTransitionTime":"2025-12-27T07:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.016585 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f11f562a-9201-4bd4-9f28-a9b2d5cd44bd-host\") pod \"node-ca-dfj4x\" (UID: \"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\") " pod="openshift-image-registry/node-ca-dfj4x" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.016619 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6bzl\" (UniqueName: \"kubernetes.io/projected/f11f562a-9201-4bd4-9f28-a9b2d5cd44bd-kube-api-access-k6bzl\") pod \"node-ca-dfj4x\" (UID: \"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\") " pod="openshift-image-registry/node-ca-dfj4x" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.016673 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f11f562a-9201-4bd4-9f28-a9b2d5cd44bd-serviceca\") pod \"node-ca-dfj4x\" (UID: \"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\") " pod="openshift-image-registry/node-ca-dfj4x" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.016677 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f11f562a-9201-4bd4-9f28-a9b2d5cd44bd-host\") pod \"node-ca-dfj4x\" (UID: \"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\") " pod="openshift-image-registry/node-ca-dfj4x" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.017481 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f11f562a-9201-4bd4-9f28-a9b2d5cd44bd-serviceca\") pod \"node-ca-dfj4x\" (UID: \"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\") " pod="openshift-image-registry/node-ca-dfj4x" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.034710 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6bzl\" (UniqueName: \"kubernetes.io/projected/f11f562a-9201-4bd4-9f28-a9b2d5cd44bd-kube-api-access-k6bzl\") pod \"node-ca-dfj4x\" (UID: \"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\") " pod="openshift-image-registry/node-ca-dfj4x" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.106466 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.106503 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.106510 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.106524 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.106535 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:42Z","lastTransitionTime":"2025-12-27T07:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.113757 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-dfj4x" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.208962 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.209018 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.209035 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.209059 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.209076 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:42Z","lastTransitionTime":"2025-12-27T07:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.311350 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.311841 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.311856 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.311872 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.311884 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:42Z","lastTransitionTime":"2025-12-27T07:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.415019 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.415052 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.415061 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.415078 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.415116 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:42Z","lastTransitionTime":"2025-12-27T07:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.517683 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.517742 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.517759 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.517790 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.517809 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:42Z","lastTransitionTime":"2025-12-27T07:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.619896 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.619941 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.619958 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.619979 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.619996 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:42Z","lastTransitionTime":"2025-12-27T07:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.621506 4934 generic.go:334] "Generic (PLEG): container finished" podID="e59598db-f0ef-42fb-b151-b436ac250b08" containerID="5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66" exitCode=0 Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.621594 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" event={"ID":"e59598db-f0ef-42fb-b151-b436ac250b08","Type":"ContainerDied","Data":"5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66"} Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.627892 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerStarted","Data":"e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c"} Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.627965 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerStarted","Data":"857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5"} Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.630492 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-dfj4x" event={"ID":"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd","Type":"ContainerStarted","Data":"722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b"} Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.630565 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-dfj4x" event={"ID":"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd","Type":"ContainerStarted","Data":"dd0afbf50d3dc0faf2b8dbcd17275fe4a62ca539528d892d5ce92336211974cd"} Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.645617 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.661307 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.690639 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.710650 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.723831 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.724525 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.724555 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.724564 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.724578 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.724589 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:42Z","lastTransitionTime":"2025-12-27T07:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.736581 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.748571 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.761507 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.776058 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.786972 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.799567 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.813434 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.826851 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.826893 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.826910 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.826928 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.826940 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:42Z","lastTransitionTime":"2025-12-27T07:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.838952 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.852508 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.867720 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.882158 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.896376 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.911458 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.920865 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.929966 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.930006 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.930018 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.930039 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.930056 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:42Z","lastTransitionTime":"2025-12-27T07:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.933877 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.944309 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.961478 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.973618 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:42 crc kubenswrapper[4934]: I1227 07:42:42.991020 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.006684 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.025612 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.032601 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.032638 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.032647 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.032662 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.032672 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:43Z","lastTransitionTime":"2025-12-27T07:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.135922 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.135969 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.135978 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.135995 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.136005 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:43Z","lastTransitionTime":"2025-12-27T07:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.163759 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.167320 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.174184 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.178785 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.196491 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.210656 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.223453 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.238936 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.238991 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.239004 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.239024 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.239040 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:43Z","lastTransitionTime":"2025-12-27T07:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.243058 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.261264 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.273326 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.285693 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.299702 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.316839 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.330561 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.341549 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.341601 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.341612 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.341631 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.341644 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:43Z","lastTransitionTime":"2025-12-27T07:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.348845 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.366976 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.384627 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.419563 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.444056 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.444194 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.444207 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.444223 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.444234 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:43Z","lastTransitionTime":"2025-12-27T07:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.461060 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.467333 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:43 crc kubenswrapper[4934]: E1227 07:42:43.467471 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.467621 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:42:43 crc kubenswrapper[4934]: E1227 07:42:43.467877 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.467888 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:42:43 crc kubenswrapper[4934]: E1227 07:42:43.468291 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.502841 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.543977 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.546794 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.546821 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.546831 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.546846 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.546861 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:43Z","lastTransitionTime":"2025-12-27T07:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.578816 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.620360 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.639491 4934 generic.go:334] "Generic (PLEG): container finished" podID="e59598db-f0ef-42fb-b151-b436ac250b08" containerID="667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874" exitCode=0 Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.639623 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" event={"ID":"e59598db-f0ef-42fb-b151-b436ac250b08","Type":"ContainerDied","Data":"667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874"} Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.649701 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.649747 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.649761 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.649780 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.649792 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:43Z","lastTransitionTime":"2025-12-27T07:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.667737 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.704569 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.742701 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.752495 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.752531 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.752542 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.752559 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.752571 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:43Z","lastTransitionTime":"2025-12-27T07:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.779052 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.821152 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.855160 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.855204 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.855216 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.855234 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.855247 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:43Z","lastTransitionTime":"2025-12-27T07:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.862806 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.907021 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.940249 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.957965 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.958007 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.958018 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.958033 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.958045 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:43Z","lastTransitionTime":"2025-12-27T07:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:43 crc kubenswrapper[4934]: I1227 07:42:43.985194 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:43Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.026439 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.060594 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.060644 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.060659 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.060678 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.060692 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:44Z","lastTransitionTime":"2025-12-27T07:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.061866 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.098744 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.135322 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.163139 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.163208 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.163229 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.163257 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.163279 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:44Z","lastTransitionTime":"2025-12-27T07:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.181051 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.224185 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.265629 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.267134 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.267207 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.267234 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.267269 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.267292 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:44Z","lastTransitionTime":"2025-12-27T07:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.305174 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.346910 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.370024 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.370068 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.370118 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.370141 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.370156 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:44Z","lastTransitionTime":"2025-12-27T07:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.388130 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.425049 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.472699 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.473743 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.473776 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.473787 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.473806 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.473825 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:44Z","lastTransitionTime":"2025-12-27T07:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.576629 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.576697 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.576723 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.576755 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.576778 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:44Z","lastTransitionTime":"2025-12-27T07:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.648288 4934 generic.go:334] "Generic (PLEG): container finished" podID="e59598db-f0ef-42fb-b151-b436ac250b08" containerID="d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d" exitCode=0 Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.648353 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" event={"ID":"e59598db-f0ef-42fb-b151-b436ac250b08","Type":"ContainerDied","Data":"d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d"} Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.658437 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerStarted","Data":"d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f"} Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.669380 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.679583 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.679640 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.679658 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.679684 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.679705 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:44Z","lastTransitionTime":"2025-12-27T07:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.690611 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.717725 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.732821 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.746659 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.765275 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.779137 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.781748 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.781792 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.781803 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.781819 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.781834 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:44Z","lastTransitionTime":"2025-12-27T07:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.791173 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.817167 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.856948 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.884619 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.884659 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.884671 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.884690 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.884702 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:44Z","lastTransitionTime":"2025-12-27T07:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.903218 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.942075 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.983794 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:44Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.987910 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.987980 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.987998 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.988019 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:44 crc kubenswrapper[4934]: I1227 07:42:44.988034 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:44Z","lastTransitionTime":"2025-12-27T07:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.019439 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:45Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.090758 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.090813 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.090830 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.090855 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.090875 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:45Z","lastTransitionTime":"2025-12-27T07:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.145700 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.145896 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.145948 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:45 crc kubenswrapper[4934]: E1227 07:42:45.145987 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:42:53.145953239 +0000 UTC m=+33.966393843 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:42:45 crc kubenswrapper[4934]: E1227 07:42:45.146133 4934 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 27 07:42:45 crc kubenswrapper[4934]: E1227 07:42:45.146151 4934 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 27 07:42:45 crc kubenswrapper[4934]: E1227 07:42:45.146222 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:53.146198266 +0000 UTC m=+33.966638900 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 27 07:42:45 crc kubenswrapper[4934]: E1227 07:42:45.146265 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:53.146237067 +0000 UTC m=+33.966677701 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.194307 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.194361 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.194380 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.194402 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.194421 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:45Z","lastTransitionTime":"2025-12-27T07:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.246895 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.246971 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:42:45 crc kubenswrapper[4934]: E1227 07:42:45.247128 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 27 07:42:45 crc kubenswrapper[4934]: E1227 07:42:45.247160 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 27 07:42:45 crc kubenswrapper[4934]: E1227 07:42:45.247171 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 27 07:42:45 crc kubenswrapper[4934]: E1227 07:42:45.247186 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 27 07:42:45 crc kubenswrapper[4934]: E1227 07:42:45.247198 4934 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:45 crc kubenswrapper[4934]: E1227 07:42:45.247203 4934 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:45 crc kubenswrapper[4934]: E1227 07:42:45.247271 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:53.247249584 +0000 UTC m=+34.067690218 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:45 crc kubenswrapper[4934]: E1227 07:42:45.247299 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:53.247284374 +0000 UTC m=+34.067724998 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.297563 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.297619 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.297638 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.297661 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.297678 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:45Z","lastTransitionTime":"2025-12-27T07:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.400967 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.401035 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.401055 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.401125 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.401152 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:45Z","lastTransitionTime":"2025-12-27T07:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.466881 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.466991 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.466991 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:45 crc kubenswrapper[4934]: E1227 07:42:45.467188 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:42:45 crc kubenswrapper[4934]: E1227 07:42:45.467524 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:42:45 crc kubenswrapper[4934]: E1227 07:42:45.467614 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.504290 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.504350 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.504371 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.504398 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.504417 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:45Z","lastTransitionTime":"2025-12-27T07:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.608891 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.608940 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.608953 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.608974 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.608990 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:45Z","lastTransitionTime":"2025-12-27T07:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.667652 4934 generic.go:334] "Generic (PLEG): container finished" podID="e59598db-f0ef-42fb-b151-b436ac250b08" containerID="b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db" exitCode=0 Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.667719 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" event={"ID":"e59598db-f0ef-42fb-b151-b436ac250b08","Type":"ContainerDied","Data":"b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db"} Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.687743 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:45Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.706500 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:45Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.712678 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.712712 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.712721 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.712736 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.712746 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:45Z","lastTransitionTime":"2025-12-27T07:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.728032 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:45Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.751737 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:45Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.766632 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:45Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.779724 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:45Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.792752 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:45Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.812356 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:45Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.814909 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.814947 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.814961 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.814982 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.814998 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:45Z","lastTransitionTime":"2025-12-27T07:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.836259 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:45Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.848754 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:45Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.864672 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:45Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.885030 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:45Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.901935 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:45Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.917166 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.917229 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.917249 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.917281 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.917300 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:45Z","lastTransitionTime":"2025-12-27T07:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:45 crc kubenswrapper[4934]: I1227 07:42:45.918133 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:45Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.019812 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.020276 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.020299 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.020323 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.020340 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:46Z","lastTransitionTime":"2025-12-27T07:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.123479 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.123551 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.123575 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.123603 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.123629 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:46Z","lastTransitionTime":"2025-12-27T07:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.226521 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.226569 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.226585 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.226607 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.226624 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:46Z","lastTransitionTime":"2025-12-27T07:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.329555 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.329624 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.329642 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.329666 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.329686 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:46Z","lastTransitionTime":"2025-12-27T07:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.432993 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.433048 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.433065 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.433120 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.433138 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:46Z","lastTransitionTime":"2025-12-27T07:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.536970 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.537023 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.537036 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.537057 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.537072 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:46Z","lastTransitionTime":"2025-12-27T07:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.639697 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.639738 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.639751 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.639768 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.639779 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:46Z","lastTransitionTime":"2025-12-27T07:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.725451 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" event={"ID":"e59598db-f0ef-42fb-b151-b436ac250b08","Type":"ContainerStarted","Data":"1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8"} Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.729279 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerStarted","Data":"c90c6ba0f7fd816d5f2072969e6e7a6c0998c96f2d59fbfc65004f53984faace"} Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.729673 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.741812 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.741846 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.741855 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.741869 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.741877 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:46Z","lastTransitionTime":"2025-12-27T07:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.742814 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:46Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.756618 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:46Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.765920 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.767776 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:46Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.780387 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:46Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.792936 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:46Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.809736 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:46Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.822138 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:46Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.836790 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:46Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.844034 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.844074 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.844111 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.844129 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.844138 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:46Z","lastTransitionTime":"2025-12-27T07:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.850802 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:46Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.864871 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:46Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.878726 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:46Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.891922 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:46Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.906639 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:46Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.922716 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:46Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.936583 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:46Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.947364 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.947439 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.947463 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.947494 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.947519 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:46Z","lastTransitionTime":"2025-12-27T07:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.954500 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:46Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:46 crc kubenswrapper[4934]: I1227 07:42:46.978916 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:46Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.001157 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:46Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.021128 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.035342 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.050541 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.050598 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.050617 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.050642 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.050660 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:47Z","lastTransitionTime":"2025-12-27T07:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.051740 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.075447 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.094139 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.114487 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.125630 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.139412 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.153414 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.154285 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.154343 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.154360 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.154386 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.154404 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:47Z","lastTransitionTime":"2025-12-27T07:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.186294 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c90c6ba0f7fd816d5f2072969e6e7a6c0998c96f2d59fbfc65004f53984faace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.257813 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.257868 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.257887 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.257911 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.257928 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:47Z","lastTransitionTime":"2025-12-27T07:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.360555 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.360619 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.360636 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.360675 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.360692 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:47Z","lastTransitionTime":"2025-12-27T07:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.463610 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.463647 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.463656 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.463670 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.463680 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:47Z","lastTransitionTime":"2025-12-27T07:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.467227 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.467272 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:42:47 crc kubenswrapper[4934]: E1227 07:42:47.467343 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:42:47 crc kubenswrapper[4934]: E1227 07:42:47.467456 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.467573 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:42:47 crc kubenswrapper[4934]: E1227 07:42:47.467720 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.566444 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.566517 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.566535 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.566561 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.566579 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:47Z","lastTransitionTime":"2025-12-27T07:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.669958 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.670021 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.670038 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.670063 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.670109 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:47Z","lastTransitionTime":"2025-12-27T07:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.732114 4934 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.735485 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.765001 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.772067 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.772136 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.772151 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.772174 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.772196 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:47Z","lastTransitionTime":"2025-12-27T07:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.779165 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.798861 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.813174 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.831175 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.845998 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.875291 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.875367 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.875389 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.875417 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.875435 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:47Z","lastTransitionTime":"2025-12-27T07:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.876688 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c90c6ba0f7fd816d5f2072969e6e7a6c0998c96f2d59fbfc65004f53984faace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.897062 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.916210 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.940412 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.960921 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.977541 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.978713 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.978905 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.979107 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.979256 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.979405 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:47Z","lastTransitionTime":"2025-12-27T07:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:47 crc kubenswrapper[4934]: I1227 07:42:47.996904 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:47Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.017182 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:48Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.036363 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:48Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.082953 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.083029 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.083057 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.083120 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.083146 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:48Z","lastTransitionTime":"2025-12-27T07:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.186296 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.186353 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.186369 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.186389 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.186406 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:48Z","lastTransitionTime":"2025-12-27T07:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.289809 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.289862 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.289880 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.289902 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.289921 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:48Z","lastTransitionTime":"2025-12-27T07:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.392215 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.392272 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.392292 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.392317 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.392334 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:48Z","lastTransitionTime":"2025-12-27T07:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.495278 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.495495 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.495575 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.495657 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.495730 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:48Z","lastTransitionTime":"2025-12-27T07:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.598324 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.598352 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.598380 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.598393 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.598402 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:48Z","lastTransitionTime":"2025-12-27T07:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.701757 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.701804 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.701852 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.701871 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.701883 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:48Z","lastTransitionTime":"2025-12-27T07:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.735595 4934 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.804188 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.804225 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.804235 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.804252 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.804264 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:48Z","lastTransitionTime":"2025-12-27T07:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.906833 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.906878 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.906894 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.906909 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:48 crc kubenswrapper[4934]: I1227 07:42:48.906920 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:48Z","lastTransitionTime":"2025-12-27T07:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.009400 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.009446 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.009457 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.009473 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.009483 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:49Z","lastTransitionTime":"2025-12-27T07:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.112135 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.112190 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.112208 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.112235 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.112260 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:49Z","lastTransitionTime":"2025-12-27T07:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.215333 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.215414 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.215433 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.215461 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.215518 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:49Z","lastTransitionTime":"2025-12-27T07:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.319202 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.319269 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.319286 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.319311 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.319331 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:49Z","lastTransitionTime":"2025-12-27T07:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.422789 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.422866 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.422912 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.422934 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.422963 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:49Z","lastTransitionTime":"2025-12-27T07:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.466594 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:42:49 crc kubenswrapper[4934]: E1227 07:42:49.466783 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.467319 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.467441 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:42:49 crc kubenswrapper[4934]: E1227 07:42:49.467564 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:42:49 crc kubenswrapper[4934]: E1227 07:42:49.467766 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.494987 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.511562 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.526591 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.526670 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.526697 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.526727 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.526751 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:49Z","lastTransitionTime":"2025-12-27T07:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.535459 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.556175 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.576721 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.595289 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.608633 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.621865 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.629793 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.629845 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.629862 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.629887 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.629904 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:49Z","lastTransitionTime":"2025-12-27T07:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.635352 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.648992 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.663578 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.680827 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.696633 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.725732 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c90c6ba0f7fd816d5f2072969e6e7a6c0998c96f2d59fbfc65004f53984faace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.732380 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.732421 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.732432 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.732450 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.732462 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:49Z","lastTransitionTime":"2025-12-27T07:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.740736 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovnkube-controller/0.log" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.745228 4934 generic.go:334] "Generic (PLEG): container finished" podID="1c4780a7-56b7-4772-af66-045a03b31a71" containerID="c90c6ba0f7fd816d5f2072969e6e7a6c0998c96f2d59fbfc65004f53984faace" exitCode=1 Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.745267 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerDied","Data":"c90c6ba0f7fd816d5f2072969e6e7a6c0998c96f2d59fbfc65004f53984faace"} Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.746470 4934 scope.go:117] "RemoveContainer" containerID="c90c6ba0f7fd816d5f2072969e6e7a6c0998c96f2d59fbfc65004f53984faace" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.764310 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.780114 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.812321 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c90c6ba0f7fd816d5f2072969e6e7a6c0998c96f2d59fbfc65004f53984faace\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c90c6ba0f7fd816d5f2072969e6e7a6c0998c96f2d59fbfc65004f53984faace\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:42:49Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1227 07:42:48.836948 6226 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1227 07:42:48.836972 6226 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1227 07:42:48.836987 6226 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1227 07:42:48.837002 6226 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1227 07:42:48.837011 6226 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1227 07:42:48.837009 6226 handler.go:208] Removed *v1.Node event handler 2\\\\nI1227 07:42:48.837025 6226 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1227 07:42:48.837028 6226 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1227 07:42:48.837041 6226 handler.go:208] Removed *v1.Node event handler 7\\\\nI1227 07:42:48.837041 6226 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1227 07:42:48.837054 6226 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1227 07:42:48.837067 6226 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1227 07:42:48.837124 6226 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1227 07:42:48.837128 6226 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1227 07:42:48.837151 6226 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.834708 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.835211 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.835241 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.835251 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.835269 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.835282 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:49Z","lastTransitionTime":"2025-12-27T07:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.853903 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.875006 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.890409 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.916538 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.932048 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.938360 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.938409 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.938429 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.938455 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.938478 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:49Z","lastTransitionTime":"2025-12-27T07:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.949596 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.970893 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:49 crc kubenswrapper[4934]: I1227 07:42:49.989075 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.004338 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:50Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.019811 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:50Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.041168 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.041216 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.041234 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.041257 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.041273 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:50Z","lastTransitionTime":"2025-12-27T07:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.144672 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.144724 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.144740 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.144763 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.144782 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:50Z","lastTransitionTime":"2025-12-27T07:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.248732 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.248789 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.248806 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.248828 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.248844 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:50Z","lastTransitionTime":"2025-12-27T07:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.352051 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.352134 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.352154 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.352178 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.352195 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:50Z","lastTransitionTime":"2025-12-27T07:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.455134 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.455177 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.455188 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.455204 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.455215 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:50Z","lastTransitionTime":"2025-12-27T07:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.557929 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.557965 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.557975 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.557989 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.557997 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:50Z","lastTransitionTime":"2025-12-27T07:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.660171 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.660218 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.660231 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.660251 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.660265 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:50Z","lastTransitionTime":"2025-12-27T07:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.752143 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovnkube-controller/0.log" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.755236 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerStarted","Data":"8da21975801894ae83e72f72da20eca3ca078a8d7e5273879881a1062849b7d1"} Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.755405 4934 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.761899 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.761953 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.761972 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.762001 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.762025 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:50Z","lastTransitionTime":"2025-12-27T07:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.774386 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:50Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.792896 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:50Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.807989 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:50Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.828115 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:50Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.843964 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:50Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.865200 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.865249 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.865263 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.865287 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.865302 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:50Z","lastTransitionTime":"2025-12-27T07:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.874940 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da21975801894ae83e72f72da20eca3ca078a8d7e5273879881a1062849b7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c90c6ba0f7fd816d5f2072969e6e7a6c0998c96f2d59fbfc65004f53984faace\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:42:49Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1227 07:42:48.836948 6226 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1227 07:42:48.836972 6226 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1227 07:42:48.836987 6226 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1227 07:42:48.837002 6226 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1227 07:42:48.837011 6226 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1227 07:42:48.837009 6226 handler.go:208] Removed *v1.Node event handler 2\\\\nI1227 07:42:48.837025 6226 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1227 07:42:48.837028 6226 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1227 07:42:48.837041 6226 handler.go:208] Removed *v1.Node event handler 7\\\\nI1227 07:42:48.837041 6226 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1227 07:42:48.837054 6226 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1227 07:42:48.837067 6226 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1227 07:42:48.837124 6226 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1227 07:42:48.837128 6226 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1227 07:42:48.837151 6226 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:50Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.890804 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:50Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.907453 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:50Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.941391 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:50Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.958316 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:50Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.968288 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.968352 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.968366 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.968385 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.968397 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:50Z","lastTransitionTime":"2025-12-27T07:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.975156 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:50Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:50 crc kubenswrapper[4934]: I1227 07:42:50.986825 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:50Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.002922 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.023223 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.051797 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.051835 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.051846 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.051864 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.051875 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:51Z","lastTransitionTime":"2025-12-27T07:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.249903 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl"] Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.251346 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.257343 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.269660 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.287768 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: E1227 07:42:51.291065 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.295445 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.295515 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.295532 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.295556 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.295573 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:51Z","lastTransitionTime":"2025-12-27T07:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.304132 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: E1227 07:42:51.319396 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.323496 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.323600 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.323671 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.323753 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.323845 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:51Z","lastTransitionTime":"2025-12-27T07:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.328775 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6jxg\" (UniqueName: \"kubernetes.io/projected/cd63c24f-fd03-4148-a9a0-a5b86882b52a-kube-api-access-v6jxg\") pod \"ovnkube-control-plane-749d76644c-g2dnl\" (UID: \"cd63c24f-fd03-4148-a9a0-a5b86882b52a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.328897 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cd63c24f-fd03-4148-a9a0-a5b86882b52a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-g2dnl\" (UID: \"cd63c24f-fd03-4148-a9a0-a5b86882b52a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.329026 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cd63c24f-fd03-4148-a9a0-a5b86882b52a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-g2dnl\" (UID: \"cd63c24f-fd03-4148-a9a0-a5b86882b52a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.329121 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cd63c24f-fd03-4148-a9a0-a5b86882b52a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-g2dnl\" (UID: \"cd63c24f-fd03-4148-a9a0-a5b86882b52a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.329440 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: E1227 07:42:51.338890 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.342449 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.342495 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.342508 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.342525 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.342537 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:51Z","lastTransitionTime":"2025-12-27T07:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.359947 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da21975801894ae83e72f72da20eca3ca078a8d7e5273879881a1062849b7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c90c6ba0f7fd816d5f2072969e6e7a6c0998c96f2d59fbfc65004f53984faace\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:42:49Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1227 07:42:48.836948 6226 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1227 07:42:48.836972 6226 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1227 07:42:48.836987 6226 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1227 07:42:48.837002 6226 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1227 07:42:48.837011 6226 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1227 07:42:48.837009 6226 handler.go:208] Removed *v1.Node event handler 2\\\\nI1227 07:42:48.837025 6226 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1227 07:42:48.837028 6226 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1227 07:42:48.837041 6226 handler.go:208] Removed *v1.Node event handler 7\\\\nI1227 07:42:48.837041 6226 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1227 07:42:48.837054 6226 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1227 07:42:48.837067 6226 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1227 07:42:48.837124 6226 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1227 07:42:48.837128 6226 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1227 07:42:48.837151 6226 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: E1227 07:42:51.362883 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.367098 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.367199 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.367256 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.367313 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.367366 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:51Z","lastTransitionTime":"2025-12-27T07:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.373347 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd63c24f-fd03-4148-a9a0-a5b86882b52a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g2dnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: E1227 07:42:51.379662 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: E1227 07:42:51.379885 4934 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.381900 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.381962 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.381987 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.382018 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.382042 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:51Z","lastTransitionTime":"2025-12-27T07:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.390889 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.406271 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.423809 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.429849 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6jxg\" (UniqueName: \"kubernetes.io/projected/cd63c24f-fd03-4148-a9a0-a5b86882b52a-kube-api-access-v6jxg\") pod \"ovnkube-control-plane-749d76644c-g2dnl\" (UID: \"cd63c24f-fd03-4148-a9a0-a5b86882b52a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.429901 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cd63c24f-fd03-4148-a9a0-a5b86882b52a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-g2dnl\" (UID: \"cd63c24f-fd03-4148-a9a0-a5b86882b52a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.429960 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cd63c24f-fd03-4148-a9a0-a5b86882b52a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-g2dnl\" (UID: \"cd63c24f-fd03-4148-a9a0-a5b86882b52a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.430027 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cd63c24f-fd03-4148-a9a0-a5b86882b52a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-g2dnl\" (UID: \"cd63c24f-fd03-4148-a9a0-a5b86882b52a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.430741 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cd63c24f-fd03-4148-a9a0-a5b86882b52a-env-overrides\") pod \"ovnkube-control-plane-749d76644c-g2dnl\" (UID: \"cd63c24f-fd03-4148-a9a0-a5b86882b52a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.431027 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cd63c24f-fd03-4148-a9a0-a5b86882b52a-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-g2dnl\" (UID: \"cd63c24f-fd03-4148-a9a0-a5b86882b52a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.436897 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cd63c24f-fd03-4148-a9a0-a5b86882b52a-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-g2dnl\" (UID: \"cd63c24f-fd03-4148-a9a0-a5b86882b52a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.444958 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.457213 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6jxg\" (UniqueName: \"kubernetes.io/projected/cd63c24f-fd03-4148-a9a0-a5b86882b52a-kube-api-access-v6jxg\") pod \"ovnkube-control-plane-749d76644c-g2dnl\" (UID: \"cd63c24f-fd03-4148-a9a0-a5b86882b52a\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.460309 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.467071 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.467157 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.467071 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:42:51 crc kubenswrapper[4934]: E1227 07:42:51.467305 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:42:51 crc kubenswrapper[4934]: E1227 07:42:51.467510 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:42:51 crc kubenswrapper[4934]: E1227 07:42:51.467625 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.480900 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.485157 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.485209 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.485221 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.485241 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.485256 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:51Z","lastTransitionTime":"2025-12-27T07:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.498221 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.518872 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.534268 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.549904 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.587116 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.588559 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.588614 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.588631 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.588654 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.588673 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:51Z","lastTransitionTime":"2025-12-27T07:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:51 crc kubenswrapper[4934]: W1227 07:42:51.608611 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd63c24f_fd03_4148_a9a0_a5b86882b52a.slice/crio-945dc4a17b3ecff7bc80dd8f50f2b2b4b7441d07215687ebc4dc50de315861ba WatchSource:0}: Error finding container 945dc4a17b3ecff7bc80dd8f50f2b2b4b7441d07215687ebc4dc50de315861ba: Status 404 returned error can't find the container with id 945dc4a17b3ecff7bc80dd8f50f2b2b4b7441d07215687ebc4dc50de315861ba Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.691760 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.691820 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.691837 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.691862 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.691888 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:51Z","lastTransitionTime":"2025-12-27T07:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.762004 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" event={"ID":"cd63c24f-fd03-4148-a9a0-a5b86882b52a","Type":"ContainerStarted","Data":"945dc4a17b3ecff7bc80dd8f50f2b2b4b7441d07215687ebc4dc50de315861ba"} Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.764723 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovnkube-controller/1.log" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.765601 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovnkube-controller/0.log" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.769376 4934 generic.go:334] "Generic (PLEG): container finished" podID="1c4780a7-56b7-4772-af66-045a03b31a71" containerID="8da21975801894ae83e72f72da20eca3ca078a8d7e5273879881a1062849b7d1" exitCode=1 Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.769447 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerDied","Data":"8da21975801894ae83e72f72da20eca3ca078a8d7e5273879881a1062849b7d1"} Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.769573 4934 scope.go:117] "RemoveContainer" containerID="c90c6ba0f7fd816d5f2072969e6e7a6c0998c96f2d59fbfc65004f53984faace" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.771245 4934 scope.go:117] "RemoveContainer" containerID="8da21975801894ae83e72f72da20eca3ca078a8d7e5273879881a1062849b7d1" Dec 27 07:42:51 crc kubenswrapper[4934]: E1227 07:42:51.771514 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.788070 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.798637 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.798827 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.798909 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.799011 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.799113 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:51Z","lastTransitionTime":"2025-12-27T07:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.814587 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da21975801894ae83e72f72da20eca3ca078a8d7e5273879881a1062849b7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c90c6ba0f7fd816d5f2072969e6e7a6c0998c96f2d59fbfc65004f53984faace\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:42:49Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1227 07:42:48.836948 6226 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1227 07:42:48.836972 6226 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1227 07:42:48.836987 6226 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1227 07:42:48.837002 6226 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1227 07:42:48.837011 6226 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1227 07:42:48.837009 6226 handler.go:208] Removed *v1.Node event handler 2\\\\nI1227 07:42:48.837025 6226 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1227 07:42:48.837028 6226 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1227 07:42:48.837041 6226 handler.go:208] Removed *v1.Node event handler 7\\\\nI1227 07:42:48.837041 6226 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1227 07:42:48.837054 6226 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1227 07:42:48.837067 6226 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1227 07:42:48.837124 6226 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1227 07:42:48.837128 6226 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1227 07:42:48.837151 6226 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8da21975801894ae83e72f72da20eca3ca078a8d7e5273879881a1062849b7d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"emoved *v1.Pod event handler 3\\\\nI1227 07:42:50.925504 6349 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.119:443: 10.217.5.119:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d4efc4a8-c514-4a6b-901c-2953978b50d3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1227 07:42:50.925567 6349 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1227 07:42:50.925631 6349 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1227 07:42:50.925662 6349 ovnkube.go:599] Stopped ovnkube\\\\nI1227 07:42:50.925681 6349 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1227 07:42:50.925741 6349 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.827611 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd63c24f-fd03-4148-a9a0-a5b86882b52a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g2dnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.842885 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.859060 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.878030 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.899492 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.902430 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.902496 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.902514 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.902539 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.902560 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:51Z","lastTransitionTime":"2025-12-27T07:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.914949 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.932947 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.951128 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.966241 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.981070 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:51 crc kubenswrapper[4934]: I1227 07:42:51.996830 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:51Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.004983 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.005023 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.005034 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.005051 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.005062 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:52Z","lastTransitionTime":"2025-12-27T07:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.012952 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:52Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.030251 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:52Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.107568 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.107608 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.107620 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.107636 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.107650 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:52Z","lastTransitionTime":"2025-12-27T07:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.210480 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.210533 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.210550 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.210574 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.210593 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:52Z","lastTransitionTime":"2025-12-27T07:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.313816 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.313878 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.313893 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.313916 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.313937 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:52Z","lastTransitionTime":"2025-12-27T07:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.417235 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.417332 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.417351 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.417380 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.417399 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:52Z","lastTransitionTime":"2025-12-27T07:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.520922 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.520986 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.521005 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.521031 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.521052 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:52Z","lastTransitionTime":"2025-12-27T07:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.624836 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.624903 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.624928 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.624953 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.624972 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:52Z","lastTransitionTime":"2025-12-27T07:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.727667 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.728182 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.728197 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.728224 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.728240 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:52Z","lastTransitionTime":"2025-12-27T07:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.740149 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-jjlqg"] Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.740924 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:42:52 crc kubenswrapper[4934]: E1227 07:42:52.741040 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.763280 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:52Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.782956 4934 scope.go:117] "RemoveContainer" containerID="8da21975801894ae83e72f72da20eca3ca078a8d7e5273879881a1062849b7d1" Dec 27 07:42:52 crc kubenswrapper[4934]: E1227 07:42:52.783271 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.784713 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:52Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.799904 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:52Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.814684 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"304709d3-05ab-4d27-8722-f94053159a29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jjlqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:52Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.830899 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.830931 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.830945 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.830960 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.830973 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:52Z","lastTransitionTime":"2025-12-27T07:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.833922 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:52Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.846199 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs\") pod \"network-metrics-daemon-jjlqg\" (UID: \"304709d3-05ab-4d27-8722-f94053159a29\") " pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.846239 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4hwf\" (UniqueName: \"kubernetes.io/projected/304709d3-05ab-4d27-8722-f94053159a29-kube-api-access-n4hwf\") pod \"network-metrics-daemon-jjlqg\" (UID: \"304709d3-05ab-4d27-8722-f94053159a29\") " pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.851675 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:52Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.880623 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da21975801894ae83e72f72da20eca3ca078a8d7e5273879881a1062849b7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c90c6ba0f7fd816d5f2072969e6e7a6c0998c96f2d59fbfc65004f53984faace\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:42:49Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1227 07:42:48.836948 6226 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1227 07:42:48.836972 6226 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1227 07:42:48.836987 6226 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1227 07:42:48.837002 6226 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1227 07:42:48.837011 6226 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1227 07:42:48.837009 6226 handler.go:208] Removed *v1.Node event handler 2\\\\nI1227 07:42:48.837025 6226 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1227 07:42:48.837028 6226 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1227 07:42:48.837041 6226 handler.go:208] Removed *v1.Node event handler 7\\\\nI1227 07:42:48.837041 6226 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1227 07:42:48.837054 6226 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1227 07:42:48.837067 6226 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1227 07:42:48.837124 6226 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1227 07:42:48.837128 6226 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1227 07:42:48.837151 6226 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8da21975801894ae83e72f72da20eca3ca078a8d7e5273879881a1062849b7d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"emoved *v1.Pod event handler 3\\\\nI1227 07:42:50.925504 6349 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.119:443: 10.217.5.119:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d4efc4a8-c514-4a6b-901c-2953978b50d3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1227 07:42:50.925567 6349 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1227 07:42:50.925631 6349 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1227 07:42:50.925662 6349 ovnkube.go:599] Stopped ovnkube\\\\nI1227 07:42:50.925681 6349 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1227 07:42:50.925741 6349 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:52Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.896546 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd63c24f-fd03-4148-a9a0-a5b86882b52a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g2dnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:52Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.909270 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:52Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.926280 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:52Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.937707 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.937772 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.937795 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.937828 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.937851 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:52Z","lastTransitionTime":"2025-12-27T07:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.946992 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs\") pod \"network-metrics-daemon-jjlqg\" (UID: \"304709d3-05ab-4d27-8722-f94053159a29\") " pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.947039 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4hwf\" (UniqueName: \"kubernetes.io/projected/304709d3-05ab-4d27-8722-f94053159a29-kube-api-access-n4hwf\") pod \"network-metrics-daemon-jjlqg\" (UID: \"304709d3-05ab-4d27-8722-f94053159a29\") " pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:42:52 crc kubenswrapper[4934]: E1227 07:42:52.947258 4934 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 27 07:42:52 crc kubenswrapper[4934]: E1227 07:42:52.947360 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs podName:304709d3-05ab-4d27-8722-f94053159a29 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:53.447328 +0000 UTC m=+34.267768644 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs") pod "network-metrics-daemon-jjlqg" (UID: "304709d3-05ab-4d27-8722-f94053159a29") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.962260 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:52Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.974282 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4hwf\" (UniqueName: \"kubernetes.io/projected/304709d3-05ab-4d27-8722-f94053159a29-kube-api-access-n4hwf\") pod \"network-metrics-daemon-jjlqg\" (UID: \"304709d3-05ab-4d27-8722-f94053159a29\") " pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:42:52 crc kubenswrapper[4934]: I1227 07:42:52.987504 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:52Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.002790 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.013545 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.024048 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.036836 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.040732 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.040775 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.040792 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.040814 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.040830 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:53Z","lastTransitionTime":"2025-12-27T07:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.050662 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.067817 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.085306 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.099012 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"304709d3-05ab-4d27-8722-f94053159a29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jjlqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.112993 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.124489 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.144007 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.144075 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.144118 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.144135 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.144147 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:53Z","lastTransitionTime":"2025-12-27T07:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.148169 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da21975801894ae83e72f72da20eca3ca078a8d7e5273879881a1062849b7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8da21975801894ae83e72f72da20eca3ca078a8d7e5273879881a1062849b7d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"emoved *v1.Pod event handler 3\\\\nI1227 07:42:50.925504 6349 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.119:443: 10.217.5.119:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d4efc4a8-c514-4a6b-901c-2953978b50d3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1227 07:42:50.925567 6349 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1227 07:42:50.925631 6349 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1227 07:42:50.925662 6349 ovnkube.go:599] Stopped ovnkube\\\\nI1227 07:42:50.925681 6349 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1227 07:42:50.925741 6349 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.148942 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.149120 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:53 crc kubenswrapper[4934]: E1227 07:42:53.149175 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:43:09.149144718 +0000 UTC m=+49.969585352 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:42:53 crc kubenswrapper[4934]: E1227 07:42:53.149232 4934 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.149271 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:53 crc kubenswrapper[4934]: E1227 07:42:53.149302 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-27 07:43:09.149281231 +0000 UTC m=+49.969721865 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 27 07:42:53 crc kubenswrapper[4934]: E1227 07:42:53.149415 4934 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 27 07:42:53 crc kubenswrapper[4934]: E1227 07:42:53.149483 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-27 07:43:09.149465866 +0000 UTC m=+49.969906500 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.164711 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd63c24f-fd03-4148-a9a0-a5b86882b52a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g2dnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.181884 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.200340 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.217811 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.240581 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.247373 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.247418 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.247433 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.247454 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.247466 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:53Z","lastTransitionTime":"2025-12-27T07:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.250363 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:42:53 crc kubenswrapper[4934]: E1227 07:42:53.250570 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 27 07:42:53 crc kubenswrapper[4934]: E1227 07:42:53.250631 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 27 07:42:53 crc kubenswrapper[4934]: E1227 07:42:53.250660 4934 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.250651 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:42:53 crc kubenswrapper[4934]: E1227 07:42:53.250752 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-27 07:43:09.25072108 +0000 UTC m=+50.071161754 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:53 crc kubenswrapper[4934]: E1227 07:42:53.250854 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 27 07:42:53 crc kubenswrapper[4934]: E1227 07:42:53.250880 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 27 07:42:53 crc kubenswrapper[4934]: E1227 07:42:53.250895 4934 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:53 crc kubenswrapper[4934]: E1227 07:42:53.250947 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-27 07:43:09.250928345 +0000 UTC m=+50.071368959 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.257160 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.270628 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.289224 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.310560 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.350316 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.350390 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.350413 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.350444 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.350465 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:53Z","lastTransitionTime":"2025-12-27T07:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.452201 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs\") pod \"network-metrics-daemon-jjlqg\" (UID: \"304709d3-05ab-4d27-8722-f94053159a29\") " pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:42:53 crc kubenswrapper[4934]: E1227 07:42:53.452374 4934 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 27 07:42:53 crc kubenswrapper[4934]: E1227 07:42:53.452473 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs podName:304709d3-05ab-4d27-8722-f94053159a29 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:54.452442505 +0000 UTC m=+35.272883139 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs") pod "network-metrics-daemon-jjlqg" (UID: "304709d3-05ab-4d27-8722-f94053159a29") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.453440 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.453500 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.453518 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.453546 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.453566 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:53Z","lastTransitionTime":"2025-12-27T07:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.467369 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.467440 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.467439 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:53 crc kubenswrapper[4934]: E1227 07:42:53.467545 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:42:53 crc kubenswrapper[4934]: E1227 07:42:53.467686 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:42:53 crc kubenswrapper[4934]: E1227 07:42:53.467993 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.556037 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.556126 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.556150 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.556181 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.556209 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:53Z","lastTransitionTime":"2025-12-27T07:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.659387 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.659611 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.659712 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.659833 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.659933 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:53Z","lastTransitionTime":"2025-12-27T07:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.762507 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.762570 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.762587 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.762611 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.762629 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:53Z","lastTransitionTime":"2025-12-27T07:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.787161 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovnkube-controller/1.log" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.792326 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" event={"ID":"cd63c24f-fd03-4148-a9a0-a5b86882b52a","Type":"ContainerStarted","Data":"6a5ed5db5f0c41fa30232392526be2a603c1fdc408aef12e8296deea70899805"} Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.792401 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" event={"ID":"cd63c24f-fd03-4148-a9a0-a5b86882b52a","Type":"ContainerStarted","Data":"bbe31fde827f2497b12876938f30cc22d79ece0e3e2ab57a034266de29d28266"} Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.816006 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.829834 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.845780 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.863150 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.866048 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.866144 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.866165 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.866190 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.866208 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:53Z","lastTransitionTime":"2025-12-27T07:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.879986 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.894970 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.906412 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.924470 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.937225 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.954404 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.966786 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.968133 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.968159 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.968166 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.968180 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.968190 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:53Z","lastTransitionTime":"2025-12-27T07:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.977192 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"304709d3-05ab-4d27-8722-f94053159a29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jjlqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:53 crc kubenswrapper[4934]: I1227 07:42:53.988807 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:53.999935 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.023127 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da21975801894ae83e72f72da20eca3ca078a8d7e5273879881a1062849b7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8da21975801894ae83e72f72da20eca3ca078a8d7e5273879881a1062849b7d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"emoved *v1.Pod event handler 3\\\\nI1227 07:42:50.925504 6349 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.119:443: 10.217.5.119:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d4efc4a8-c514-4a6b-901c-2953978b50d3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1227 07:42:50.925567 6349 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1227 07:42:50.925631 6349 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1227 07:42:50.925662 6349 ovnkube.go:599] Stopped ovnkube\\\\nI1227 07:42:50.925681 6349 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1227 07:42:50.925741 6349 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:54Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.038595 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd63c24f-fd03-4148-a9a0-a5b86882b52a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe31fde827f2497b12876938f30cc22d79ece0e3e2ab57a034266de29d28266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a5ed5db5f0c41fa30232392526be2a603c1fdc408aef12e8296deea70899805\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g2dnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:54Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.070740 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.070797 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.070808 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.070826 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.070838 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:54Z","lastTransitionTime":"2025-12-27T07:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.173774 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.173813 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.173825 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.173840 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.173851 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:54Z","lastTransitionTime":"2025-12-27T07:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.277308 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.277371 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.277384 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.277408 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.277424 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:54Z","lastTransitionTime":"2025-12-27T07:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.380306 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.380358 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.380392 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.380418 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.380435 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:54Z","lastTransitionTime":"2025-12-27T07:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.466702 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:42:54 crc kubenswrapper[4934]: E1227 07:42:54.466913 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.467568 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs\") pod \"network-metrics-daemon-jjlqg\" (UID: \"304709d3-05ab-4d27-8722-f94053159a29\") " pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:42:54 crc kubenswrapper[4934]: E1227 07:42:54.467793 4934 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 27 07:42:54 crc kubenswrapper[4934]: E1227 07:42:54.467881 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs podName:304709d3-05ab-4d27-8722-f94053159a29 nodeName:}" failed. No retries permitted until 2025-12-27 07:42:56.467859354 +0000 UTC m=+37.288299978 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs") pod "network-metrics-daemon-jjlqg" (UID: "304709d3-05ab-4d27-8722-f94053159a29") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.483416 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.483467 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.483489 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.483513 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.483530 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:54Z","lastTransitionTime":"2025-12-27T07:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.586953 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.587040 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.587059 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.587120 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.587143 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:54Z","lastTransitionTime":"2025-12-27T07:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.690460 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.690536 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.690558 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.690589 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.690610 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:54Z","lastTransitionTime":"2025-12-27T07:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.794269 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.794324 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.794342 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.794369 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.794389 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:54Z","lastTransitionTime":"2025-12-27T07:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.897415 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.897475 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.897493 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.897517 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:54 crc kubenswrapper[4934]: I1227 07:42:54.897535 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:54Z","lastTransitionTime":"2025-12-27T07:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.000983 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.001038 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.001054 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.001077 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.001124 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:55Z","lastTransitionTime":"2025-12-27T07:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.104245 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.104309 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.104327 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.104352 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.104374 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:55Z","lastTransitionTime":"2025-12-27T07:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.207382 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.207453 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.207473 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.207503 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.207530 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:55Z","lastTransitionTime":"2025-12-27T07:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.310554 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.310629 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.310647 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.310672 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.310690 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:55Z","lastTransitionTime":"2025-12-27T07:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.413560 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.413619 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.413636 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.413660 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.413678 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:55Z","lastTransitionTime":"2025-12-27T07:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.466994 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.467020 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.467142 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:42:55 crc kubenswrapper[4934]: E1227 07:42:55.467803 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:42:55 crc kubenswrapper[4934]: E1227 07:42:55.467944 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:42:55 crc kubenswrapper[4934]: E1227 07:42:55.468144 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.516569 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.516604 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.516615 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.516634 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.516644 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:55Z","lastTransitionTime":"2025-12-27T07:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.619554 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.619721 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.619741 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.619765 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.619784 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:55Z","lastTransitionTime":"2025-12-27T07:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.723402 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.723462 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.723482 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.723505 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.723525 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:55Z","lastTransitionTime":"2025-12-27T07:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.826732 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.826778 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.826792 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.826814 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.826827 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:55Z","lastTransitionTime":"2025-12-27T07:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.929519 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.929611 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.929630 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.929654 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:55 crc kubenswrapper[4934]: I1227 07:42:55.929671 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:55Z","lastTransitionTime":"2025-12-27T07:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.032939 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.033025 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.033057 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.033125 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.033151 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:56Z","lastTransitionTime":"2025-12-27T07:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.135712 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.135753 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.135764 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.135779 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.135790 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:56Z","lastTransitionTime":"2025-12-27T07:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.238780 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.238841 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.238858 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.238885 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.238902 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:56Z","lastTransitionTime":"2025-12-27T07:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.342322 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.342453 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.342526 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.342559 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.342619 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:56Z","lastTransitionTime":"2025-12-27T07:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.446293 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.446372 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.446390 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.446414 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.446433 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:56Z","lastTransitionTime":"2025-12-27T07:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.466495 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:42:56 crc kubenswrapper[4934]: E1227 07:42:56.466792 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.491532 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs\") pod \"network-metrics-daemon-jjlqg\" (UID: \"304709d3-05ab-4d27-8722-f94053159a29\") " pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:42:56 crc kubenswrapper[4934]: E1227 07:42:56.491779 4934 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 27 07:42:56 crc kubenswrapper[4934]: E1227 07:42:56.491900 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs podName:304709d3-05ab-4d27-8722-f94053159a29 nodeName:}" failed. No retries permitted until 2025-12-27 07:43:00.491868488 +0000 UTC m=+41.312309122 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs") pod "network-metrics-daemon-jjlqg" (UID: "304709d3-05ab-4d27-8722-f94053159a29") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.548555 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.548626 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.548647 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.548679 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.548701 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:56Z","lastTransitionTime":"2025-12-27T07:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.651048 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.651140 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.651159 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.651184 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.651203 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:56Z","lastTransitionTime":"2025-12-27T07:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.754078 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.754164 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.754180 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.754202 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.754219 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:56Z","lastTransitionTime":"2025-12-27T07:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.856909 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.856970 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.856990 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.857016 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.857033 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:56Z","lastTransitionTime":"2025-12-27T07:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.960250 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.960294 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.960305 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.960322 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:56 crc kubenswrapper[4934]: I1227 07:42:56.960334 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:56Z","lastTransitionTime":"2025-12-27T07:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.063947 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.064020 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.064044 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.064121 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.064148 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:57Z","lastTransitionTime":"2025-12-27T07:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.167471 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.167582 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.167607 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.167636 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.167658 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:57Z","lastTransitionTime":"2025-12-27T07:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.269918 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.269954 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.269961 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.269974 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.269982 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:57Z","lastTransitionTime":"2025-12-27T07:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.372454 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.372503 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.372515 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.372531 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.372542 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:57Z","lastTransitionTime":"2025-12-27T07:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.466898 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.466987 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.466911 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:57 crc kubenswrapper[4934]: E1227 07:42:57.467159 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:42:57 crc kubenswrapper[4934]: E1227 07:42:57.467260 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:42:57 crc kubenswrapper[4934]: E1227 07:42:57.467320 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.479075 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.479136 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.479147 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.479162 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.479174 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:57Z","lastTransitionTime":"2025-12-27T07:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.581741 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.581894 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.581914 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.581939 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.581961 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:57Z","lastTransitionTime":"2025-12-27T07:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.685662 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.685733 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.685755 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.685787 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.685811 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:57Z","lastTransitionTime":"2025-12-27T07:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.787710 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.787777 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.787799 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.787829 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.787853 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:57Z","lastTransitionTime":"2025-12-27T07:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.890735 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.890803 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.890828 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.890856 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.890877 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:57Z","lastTransitionTime":"2025-12-27T07:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.993861 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.993925 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.993942 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.993968 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:57 crc kubenswrapper[4934]: I1227 07:42:57.993985 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:57Z","lastTransitionTime":"2025-12-27T07:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.096847 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.096913 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.096931 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.096958 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.096979 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:58Z","lastTransitionTime":"2025-12-27T07:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.199794 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.199854 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.199873 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.199896 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.199913 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:58Z","lastTransitionTime":"2025-12-27T07:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.302655 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.302746 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.302767 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.302824 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.302854 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:58Z","lastTransitionTime":"2025-12-27T07:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.405907 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.405972 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.405990 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.406014 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.406035 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:58Z","lastTransitionTime":"2025-12-27T07:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.466967 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:42:58 crc kubenswrapper[4934]: E1227 07:42:58.467157 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.508242 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.508289 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.508303 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.508323 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.508338 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:58Z","lastTransitionTime":"2025-12-27T07:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.610545 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.610604 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.610621 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.610643 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.610659 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:58Z","lastTransitionTime":"2025-12-27T07:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.713056 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.713145 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.713164 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.713190 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.713208 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:58Z","lastTransitionTime":"2025-12-27T07:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.815981 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.816490 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.816655 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.816821 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.816975 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:58Z","lastTransitionTime":"2025-12-27T07:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.919520 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.919586 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.919597 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.919617 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:58 crc kubenswrapper[4934]: I1227 07:42:58.919649 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:58Z","lastTransitionTime":"2025-12-27T07:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.022575 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.023237 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.023285 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.023321 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.023346 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:59Z","lastTransitionTime":"2025-12-27T07:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.126317 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.126368 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.126378 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.126398 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.126410 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:59Z","lastTransitionTime":"2025-12-27T07:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.229267 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.229328 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.229343 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.229367 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.229386 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:59Z","lastTransitionTime":"2025-12-27T07:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.332900 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.332967 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.332987 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.333010 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.333029 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:59Z","lastTransitionTime":"2025-12-27T07:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.436346 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.436415 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.436435 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.436464 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.436485 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:59Z","lastTransitionTime":"2025-12-27T07:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.466542 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.466604 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:42:59 crc kubenswrapper[4934]: E1227 07:42:59.466727 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:42:59 crc kubenswrapper[4934]: E1227 07:42:59.466921 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.466575 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:42:59 crc kubenswrapper[4934]: E1227 07:42:59.467394 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.487465 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.504676 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.526506 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.541939 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.542065 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.542192 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.542289 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.542396 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:59Z","lastTransitionTime":"2025-12-27T07:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.550209 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.565735 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.590110 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.604496 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.620963 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.635028 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"304709d3-05ab-4d27-8722-f94053159a29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jjlqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.645258 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.645391 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.645475 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.645588 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.645692 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:59Z","lastTransitionTime":"2025-12-27T07:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.651682 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.668777 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.681670 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.697478 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd63c24f-fd03-4148-a9a0-a5b86882b52a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe31fde827f2497b12876938f30cc22d79ece0e3e2ab57a034266de29d28266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a5ed5db5f0c41fa30232392526be2a603c1fdc408aef12e8296deea70899805\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g2dnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.723347 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.737444 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.748472 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.748531 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.748548 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.748571 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.748589 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:59Z","lastTransitionTime":"2025-12-27T07:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.760562 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da21975801894ae83e72f72da20eca3ca078a8d7e5273879881a1062849b7d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8da21975801894ae83e72f72da20eca3ca078a8d7e5273879881a1062849b7d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"emoved *v1.Pod event handler 3\\\\nI1227 07:42:50.925504 6349 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.119:443: 10.217.5.119:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d4efc4a8-c514-4a6b-901c-2953978b50d3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1227 07:42:50.925567 6349 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1227 07:42:50.925631 6349 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1227 07:42:50.925662 6349 ovnkube.go:599] Stopped ovnkube\\\\nI1227 07:42:50.925681 6349 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1227 07:42:50.925741 6349 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:42:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.851658 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.851773 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.851797 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.851826 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.851843 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:59Z","lastTransitionTime":"2025-12-27T07:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.955055 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.955158 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.955179 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.955203 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:42:59 crc kubenswrapper[4934]: I1227 07:42:59.955220 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:42:59Z","lastTransitionTime":"2025-12-27T07:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.058273 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.058333 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.058350 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.058374 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.058394 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:00Z","lastTransitionTime":"2025-12-27T07:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.161526 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.161590 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.161615 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.161647 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.161671 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:00Z","lastTransitionTime":"2025-12-27T07:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.264699 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.264747 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.264765 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.264788 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.264806 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:00Z","lastTransitionTime":"2025-12-27T07:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.367386 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.367494 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.367525 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.367565 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.367589 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:00Z","lastTransitionTime":"2025-12-27T07:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.466684 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:00 crc kubenswrapper[4934]: E1227 07:43:00.466899 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.471307 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.471389 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.471407 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.471432 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.471450 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:00Z","lastTransitionTime":"2025-12-27T07:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.536828 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs\") pod \"network-metrics-daemon-jjlqg\" (UID: \"304709d3-05ab-4d27-8722-f94053159a29\") " pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:00 crc kubenswrapper[4934]: E1227 07:43:00.537257 4934 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 27 07:43:00 crc kubenswrapper[4934]: E1227 07:43:00.537401 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs podName:304709d3-05ab-4d27-8722-f94053159a29 nodeName:}" failed. No retries permitted until 2025-12-27 07:43:08.537358407 +0000 UTC m=+49.357799041 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs") pod "network-metrics-daemon-jjlqg" (UID: "304709d3-05ab-4d27-8722-f94053159a29") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.576689 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.576767 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.576788 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.576816 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.576838 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:00Z","lastTransitionTime":"2025-12-27T07:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.680601 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.680700 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.680729 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.680767 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.680826 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:00Z","lastTransitionTime":"2025-12-27T07:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.784569 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.784635 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.784656 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.784681 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.784699 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:00Z","lastTransitionTime":"2025-12-27T07:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.888727 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.888779 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.888793 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.888815 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.888827 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:00Z","lastTransitionTime":"2025-12-27T07:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.992913 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.992975 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.992988 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.993007 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:00 crc kubenswrapper[4934]: I1227 07:43:00.993022 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:00Z","lastTransitionTime":"2025-12-27T07:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.095897 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.096513 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.096551 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.096580 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.096601 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:01Z","lastTransitionTime":"2025-12-27T07:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.199524 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.199596 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.199614 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.199638 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.199655 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:01Z","lastTransitionTime":"2025-12-27T07:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.302714 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.302778 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.302797 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.302822 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.302841 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:01Z","lastTransitionTime":"2025-12-27T07:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.406296 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.406394 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.406415 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.406447 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.406467 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:01Z","lastTransitionTime":"2025-12-27T07:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.467058 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.467120 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:01 crc kubenswrapper[4934]: E1227 07:43:01.467350 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:01 crc kubenswrapper[4934]: E1227 07:43:01.467539 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.467120 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:01 crc kubenswrapper[4934]: E1227 07:43:01.467802 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.509181 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.509248 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.509265 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.509288 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.509307 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:01Z","lastTransitionTime":"2025-12-27T07:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.592315 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.592453 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.592473 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.592510 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.592527 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:01Z","lastTransitionTime":"2025-12-27T07:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:01 crc kubenswrapper[4934]: E1227 07:43:01.613839 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:01Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.618581 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.618636 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.618653 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.618679 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.618696 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:01Z","lastTransitionTime":"2025-12-27T07:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:01 crc kubenswrapper[4934]: E1227 07:43:01.636187 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:01Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.640366 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.640448 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.640479 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.640514 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.640538 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:01Z","lastTransitionTime":"2025-12-27T07:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:01 crc kubenswrapper[4934]: E1227 07:43:01.659756 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:01Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.663979 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.664047 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.664068 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.664130 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.664153 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:01Z","lastTransitionTime":"2025-12-27T07:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:01 crc kubenswrapper[4934]: E1227 07:43:01.685682 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:01Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.690494 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.690598 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.690615 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.690631 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.690643 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:01Z","lastTransitionTime":"2025-12-27T07:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:01 crc kubenswrapper[4934]: E1227 07:43:01.706262 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:01Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:01 crc kubenswrapper[4934]: E1227 07:43:01.706486 4934 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.708314 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.708344 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.708357 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.708373 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.708386 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:01Z","lastTransitionTime":"2025-12-27T07:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.810941 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.810993 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.811005 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.811028 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.811044 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:01Z","lastTransitionTime":"2025-12-27T07:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.914829 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.914908 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.914927 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.914956 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:01 crc kubenswrapper[4934]: I1227 07:43:01.914976 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:01Z","lastTransitionTime":"2025-12-27T07:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.018061 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.018141 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.018156 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.018182 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.018198 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:02Z","lastTransitionTime":"2025-12-27T07:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.121338 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.121428 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.121456 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.121485 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.121504 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:02Z","lastTransitionTime":"2025-12-27T07:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.225378 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.225456 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.225475 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.225501 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.225519 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:02Z","lastTransitionTime":"2025-12-27T07:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.329057 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.329167 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.329187 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.329211 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.329227 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:02Z","lastTransitionTime":"2025-12-27T07:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.432277 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.432338 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.432360 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.432388 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.432410 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:02Z","lastTransitionTime":"2025-12-27T07:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.467150 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:02 crc kubenswrapper[4934]: E1227 07:43:02.467299 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.535246 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.535316 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.535335 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.535358 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.535375 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:02Z","lastTransitionTime":"2025-12-27T07:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.638220 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.638281 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.638299 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.638324 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.638343 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:02Z","lastTransitionTime":"2025-12-27T07:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.741147 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.741204 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.741221 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.741249 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.741266 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:02Z","lastTransitionTime":"2025-12-27T07:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.844059 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.844152 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.844171 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.844194 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.844213 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:02Z","lastTransitionTime":"2025-12-27T07:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.947263 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.947364 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.947384 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.947417 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:02 crc kubenswrapper[4934]: I1227 07:43:02.947443 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:02Z","lastTransitionTime":"2025-12-27T07:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.050075 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.050187 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.050213 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.050241 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.050263 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:03Z","lastTransitionTime":"2025-12-27T07:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.153097 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.153158 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.153171 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.153187 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.153198 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:03Z","lastTransitionTime":"2025-12-27T07:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.256248 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.256297 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.256308 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.256324 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.256337 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:03Z","lastTransitionTime":"2025-12-27T07:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.358467 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.358550 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.358571 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.358602 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.358625 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:03Z","lastTransitionTime":"2025-12-27T07:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.461019 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.461120 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.461138 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.461163 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.461180 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:03Z","lastTransitionTime":"2025-12-27T07:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.467435 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.467483 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:03 crc kubenswrapper[4934]: E1227 07:43:03.467609 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.467667 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:03 crc kubenswrapper[4934]: E1227 07:43:03.468248 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:03 crc kubenswrapper[4934]: E1227 07:43:03.468314 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.468724 4934 scope.go:117] "RemoveContainer" containerID="8da21975801894ae83e72f72da20eca3ca078a8d7e5273879881a1062849b7d1" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.564396 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.564781 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.564800 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.564824 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.564841 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:03Z","lastTransitionTime":"2025-12-27T07:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.667467 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.667507 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.667518 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.667533 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.667544 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:03Z","lastTransitionTime":"2025-12-27T07:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.770732 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.770786 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.770800 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.770822 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.770836 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:03Z","lastTransitionTime":"2025-12-27T07:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.874167 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.874217 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.874228 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.874244 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.874257 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:03Z","lastTransitionTime":"2025-12-27T07:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.976777 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.976880 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.976904 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.977387 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:03 crc kubenswrapper[4934]: I1227 07:43:03.977644 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:03Z","lastTransitionTime":"2025-12-27T07:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.079555 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.079602 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.079611 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.079628 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.079640 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:04Z","lastTransitionTime":"2025-12-27T07:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.194926 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.194979 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.194995 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.195018 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.195036 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:04Z","lastTransitionTime":"2025-12-27T07:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.297307 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.297349 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.297360 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.297376 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.297388 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:04Z","lastTransitionTime":"2025-12-27T07:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.399453 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.399505 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.399520 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.399541 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.399555 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:04Z","lastTransitionTime":"2025-12-27T07:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.467104 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:04 crc kubenswrapper[4934]: E1227 07:43:04.467257 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.501888 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.501935 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.501952 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.501970 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.501982 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:04Z","lastTransitionTime":"2025-12-27T07:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.606636 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.606713 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.606732 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.606754 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.606772 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:04Z","lastTransitionTime":"2025-12-27T07:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.708972 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.709012 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.709021 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.709037 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.709047 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:04Z","lastTransitionTime":"2025-12-27T07:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.812570 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.812650 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.812669 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.812699 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.812718 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:04Z","lastTransitionTime":"2025-12-27T07:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.833166 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovnkube-controller/1.log" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.836879 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerStarted","Data":"8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05"} Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.836983 4934 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.864633 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:04Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.887572 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:04Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.910149 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:04Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.914922 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.914951 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.914959 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.914973 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.914982 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:04Z","lastTransitionTime":"2025-12-27T07:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.926588 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:04Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.936853 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:04Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.951232 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:04Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.969981 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:04Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:04 crc kubenswrapper[4934]: I1227 07:43:04.984936 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:04Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.002834 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:05Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.017000 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.017024 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.017033 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.017045 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.017053 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:05Z","lastTransitionTime":"2025-12-27T07:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.020168 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:05Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.034906 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:05Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.046975 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"304709d3-05ab-4d27-8722-f94053159a29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jjlqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:05Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.065332 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:05Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.078724 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:05Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.109842 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8da21975801894ae83e72f72da20eca3ca078a8d7e5273879881a1062849b7d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"emoved *v1.Pod event handler 3\\\\nI1227 07:42:50.925504 6349 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.119:443: 10.217.5.119:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d4efc4a8-c514-4a6b-901c-2953978b50d3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1227 07:42:50.925567 6349 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1227 07:42:50.925631 6349 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1227 07:42:50.925662 6349 ovnkube.go:599] Stopped ovnkube\\\\nI1227 07:42:50.925681 6349 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1227 07:42:50.925741 6349 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:43:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:05Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.121033 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.121072 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.121099 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.121111 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.121121 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:05Z","lastTransitionTime":"2025-12-27T07:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.126774 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd63c24f-fd03-4148-a9a0-a5b86882b52a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe31fde827f2497b12876938f30cc22d79ece0e3e2ab57a034266de29d28266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a5ed5db5f0c41fa30232392526be2a603c1fdc408aef12e8296deea70899805\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g2dnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:05Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.224484 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.224559 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.224585 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.224620 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.224644 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:05Z","lastTransitionTime":"2025-12-27T07:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.327944 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.327987 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.327997 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.328015 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.328026 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:05Z","lastTransitionTime":"2025-12-27T07:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.430583 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.430958 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.431191 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.431432 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.431632 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:05Z","lastTransitionTime":"2025-12-27T07:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.467330 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.467479 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.467498 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:05 crc kubenswrapper[4934]: E1227 07:43:05.468119 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:05 crc kubenswrapper[4934]: E1227 07:43:05.468204 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:05 crc kubenswrapper[4934]: E1227 07:43:05.468508 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.535238 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.535298 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.535313 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.535334 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.535349 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:05Z","lastTransitionTime":"2025-12-27T07:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.637671 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.637962 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.638060 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.638236 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.638420 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:05Z","lastTransitionTime":"2025-12-27T07:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.740953 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.740994 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.741005 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.741026 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.741038 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:05Z","lastTransitionTime":"2025-12-27T07:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.841126 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovnkube-controller/2.log" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.841934 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovnkube-controller/1.log" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.842635 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.842661 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.842670 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.842684 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.842696 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:05Z","lastTransitionTime":"2025-12-27T07:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.844459 4934 generic.go:334] "Generic (PLEG): container finished" podID="1c4780a7-56b7-4772-af66-045a03b31a71" containerID="8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05" exitCode=1 Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.844511 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerDied","Data":"8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05"} Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.844550 4934 scope.go:117] "RemoveContainer" containerID="8da21975801894ae83e72f72da20eca3ca078a8d7e5273879881a1062849b7d1" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.845326 4934 scope.go:117] "RemoveContainer" containerID="8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05" Dec 27 07:43:05 crc kubenswrapper[4934]: E1227 07:43:05.845490 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.862932 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:05Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.875416 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:05Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.894564 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:05Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.910289 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:05Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.930182 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:05Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.945058 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.945136 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.945154 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.945179 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.945196 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:05Z","lastTransitionTime":"2025-12-27T07:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.945370 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"304709d3-05ab-4d27-8722-f94053159a29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jjlqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:05Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.968213 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:05Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.986876 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:05Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:05 crc kubenswrapper[4934]: I1227 07:43:05.997619 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:05Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.011643 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd63c24f-fd03-4148-a9a0-a5b86882b52a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe31fde827f2497b12876938f30cc22d79ece0e3e2ab57a034266de29d28266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a5ed5db5f0c41fa30232392526be2a603c1fdc408aef12e8296deea70899805\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g2dnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:06Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.028386 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:06Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.045199 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:06Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.047225 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.047280 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.047303 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.047332 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.047354 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:06Z","lastTransitionTime":"2025-12-27T07:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.071139 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8da21975801894ae83e72f72da20eca3ca078a8d7e5273879881a1062849b7d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"message\\\":\\\"emoved *v1.Pod event handler 3\\\\nI1227 07:42:50.925504 6349 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-multus/multus-admission-controller]} name:Service_openshift-multus/multus-admission-controller_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.119:443: 10.217.5.119:8443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d4efc4a8-c514-4a6b-901c-2953978b50d3}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:NB_Global Row:map[] Rows:[] Columns:[] Mutations:[{Column:nb_cfg Mutator:+= Value:1}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {6011affd-30a6-4be6-872d-e4cf1ca780cf}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1227 07:42:50.925567 6349 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1227 07:42:50.925631 6349 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1227 07:42:50.925662 6349 ovnkube.go:599] Stopped ovnkube\\\\nI1227 07:42:50.925681 6349 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1227 07:42:50.925741 6349 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:04Z\\\",\\\"message\\\":\\\":43:04.648259 6555 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1227 07:43:04.648311 6555 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1227 07:43:04.648332 6555 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1227 07:43:04.648368 6555 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1227 07:43:04.648417 6555 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1227 07:43:04.648433 6555 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1227 07:43:04.648472 6555 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1227 07:43:04.648490 6555 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1227 07:43:04.648506 6555 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1227 07:43:04.648525 6555 handler.go:208] Removed *v1.Node event handler 2\\\\nI1227 07:43:04.648539 6555 handler.go:208] Removed *v1.Node event handler 7\\\\nI1227 07:43:04.648553 6555 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1227 07:43:04.648568 6555 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1227 07:43:04.649724 6555 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1227 07:43:04.649797 6555 factory.go:656] Stopping watch factory\\\\nI1227 07:43:04.649819 6555 ovnkube.go:599] Stopped ovnkube\\\\nI1227 07:43:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:43:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:06Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.094250 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:06Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.110636 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:06Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.127940 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:06Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.149682 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.150013 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.150209 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.150360 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.150516 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:06Z","lastTransitionTime":"2025-12-27T07:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.253914 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.254972 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.255033 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.255136 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.255157 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:06Z","lastTransitionTime":"2025-12-27T07:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.357720 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.357758 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.357771 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.357789 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.357801 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:06Z","lastTransitionTime":"2025-12-27T07:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.460970 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.461274 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.461382 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.461489 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.461631 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:06Z","lastTransitionTime":"2025-12-27T07:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.467175 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:06 crc kubenswrapper[4934]: E1227 07:43:06.467299 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.564569 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.564613 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.564624 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.564640 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.564651 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:06Z","lastTransitionTime":"2025-12-27T07:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.668721 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.668801 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.668826 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.668855 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.668881 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:06Z","lastTransitionTime":"2025-12-27T07:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.772619 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.772679 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.772699 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.772728 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.772753 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:06Z","lastTransitionTime":"2025-12-27T07:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.853277 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovnkube-controller/2.log" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.875921 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.876000 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.876026 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.876059 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.876092 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:06Z","lastTransitionTime":"2025-12-27T07:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.979786 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.980362 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.980515 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.980672 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:06 crc kubenswrapper[4934]: I1227 07:43:06.980808 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:06Z","lastTransitionTime":"2025-12-27T07:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.083521 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.083561 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.083571 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.083586 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.083599 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:07Z","lastTransitionTime":"2025-12-27T07:43:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.186229 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.186295 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.186327 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.186359 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.186384 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:07Z","lastTransitionTime":"2025-12-27T07:43:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.289031 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.289150 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.289169 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.289186 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.289229 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:07Z","lastTransitionTime":"2025-12-27T07:43:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.392185 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.392231 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.392246 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.392262 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.392275 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:07Z","lastTransitionTime":"2025-12-27T07:43:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.467188 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.467307 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.467380 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:07 crc kubenswrapper[4934]: E1227 07:43:07.467600 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:07 crc kubenswrapper[4934]: E1227 07:43:07.467770 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:07 crc kubenswrapper[4934]: E1227 07:43:07.467972 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.494678 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.494718 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.494732 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.494751 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.494766 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:07Z","lastTransitionTime":"2025-12-27T07:43:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.596720 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.596766 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.596785 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.596805 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.596818 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:07Z","lastTransitionTime":"2025-12-27T07:43:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.699355 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.699426 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.699451 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.699479 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.699500 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:07Z","lastTransitionTime":"2025-12-27T07:43:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.806531 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.806585 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.806611 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.806637 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.806656 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:07Z","lastTransitionTime":"2025-12-27T07:43:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.909434 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.909495 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.909513 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.909537 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:07 crc kubenswrapper[4934]: I1227 07:43:07.909555 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:07Z","lastTransitionTime":"2025-12-27T07:43:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.011960 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.012022 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.012042 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.012066 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.012087 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:08Z","lastTransitionTime":"2025-12-27T07:43:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.115216 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.115265 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.115282 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.115306 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.115325 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:08Z","lastTransitionTime":"2025-12-27T07:43:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.218055 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.218433 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.218656 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.218883 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.219066 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:08Z","lastTransitionTime":"2025-12-27T07:43:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.257693 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.259246 4934 scope.go:117] "RemoveContainer" containerID="8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05" Dec 27 07:43:08 crc kubenswrapper[4934]: E1227 07:43:08.259597 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.283730 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:08Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.305994 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:08Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.321967 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.322021 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.322039 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.322065 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.322136 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:08Z","lastTransitionTime":"2025-12-27T07:43:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.329476 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:08Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.348927 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:08Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.369471 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:08Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.384596 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:08Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.399561 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:08Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.418011 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:08Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.425079 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.425141 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.425157 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.425177 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.425191 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:08Z","lastTransitionTime":"2025-12-27T07:43:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.437244 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:08Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.454992 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:08Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.466924 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:08 crc kubenswrapper[4934]: E1227 07:43:08.467381 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.471279 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:08Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.487040 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"304709d3-05ab-4d27-8722-f94053159a29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jjlqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:08Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.507591 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:08Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.524639 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:08Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.527948 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.528020 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.528039 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.528063 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.528087 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:08Z","lastTransitionTime":"2025-12-27T07:43:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.545193 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs\") pod \"network-metrics-daemon-jjlqg\" (UID: \"304709d3-05ab-4d27-8722-f94053159a29\") " pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:08 crc kubenswrapper[4934]: E1227 07:43:08.545378 4934 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 27 07:43:08 crc kubenswrapper[4934]: E1227 07:43:08.545763 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs podName:304709d3-05ab-4d27-8722-f94053159a29 nodeName:}" failed. No retries permitted until 2025-12-27 07:43:24.545737114 +0000 UTC m=+65.366177748 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs") pod "network-metrics-daemon-jjlqg" (UID: "304709d3-05ab-4d27-8722-f94053159a29") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.556940 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:04Z\\\",\\\"message\\\":\\\":43:04.648259 6555 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1227 07:43:04.648311 6555 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1227 07:43:04.648332 6555 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1227 07:43:04.648368 6555 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1227 07:43:04.648417 6555 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1227 07:43:04.648433 6555 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1227 07:43:04.648472 6555 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1227 07:43:04.648490 6555 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1227 07:43:04.648506 6555 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1227 07:43:04.648525 6555 handler.go:208] Removed *v1.Node event handler 2\\\\nI1227 07:43:04.648539 6555 handler.go:208] Removed *v1.Node event handler 7\\\\nI1227 07:43:04.648553 6555 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1227 07:43:04.648568 6555 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1227 07:43:04.649724 6555 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1227 07:43:04.649797 6555 factory.go:656] Stopping watch factory\\\\nI1227 07:43:04.649819 6555 ovnkube.go:599] Stopped ovnkube\\\\nI1227 07:43:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:43:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:08Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.573652 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd63c24f-fd03-4148-a9a0-a5b86882b52a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe31fde827f2497b12876938f30cc22d79ece0e3e2ab57a034266de29d28266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a5ed5db5f0c41fa30232392526be2a603c1fdc408aef12e8296deea70899805\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g2dnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:08Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.631610 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.631676 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.631695 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.631722 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.631741 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:08Z","lastTransitionTime":"2025-12-27T07:43:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.734823 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.734922 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.734961 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.734994 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.735019 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:08Z","lastTransitionTime":"2025-12-27T07:43:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.837940 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.837977 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.837988 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.838004 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.838014 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:08Z","lastTransitionTime":"2025-12-27T07:43:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.941140 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.941208 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.941223 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.941256 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:08 crc kubenswrapper[4934]: I1227 07:43:08.941273 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:08Z","lastTransitionTime":"2025-12-27T07:43:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.044525 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.044597 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.044614 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.044641 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.044659 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:09Z","lastTransitionTime":"2025-12-27T07:43:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.147803 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.147855 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.147871 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.147888 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.147899 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:09Z","lastTransitionTime":"2025-12-27T07:43:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.152282 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:43:09 crc kubenswrapper[4934]: E1227 07:43:09.152457 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:43:41.152432729 +0000 UTC m=+81.972873343 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.152571 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.152618 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:09 crc kubenswrapper[4934]: E1227 07:43:09.152712 4934 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 27 07:43:09 crc kubenswrapper[4934]: E1227 07:43:09.152729 4934 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 27 07:43:09 crc kubenswrapper[4934]: E1227 07:43:09.152767 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-27 07:43:41.152758217 +0000 UTC m=+81.973198821 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 27 07:43:09 crc kubenswrapper[4934]: E1227 07:43:09.152783 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-27 07:43:41.152776218 +0000 UTC m=+81.973216822 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.250882 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.250939 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.250956 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.250980 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.250997 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:09Z","lastTransitionTime":"2025-12-27T07:43:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.253387 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.253459 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:09 crc kubenswrapper[4934]: E1227 07:43:09.253618 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 27 07:43:09 crc kubenswrapper[4934]: E1227 07:43:09.253650 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 27 07:43:09 crc kubenswrapper[4934]: E1227 07:43:09.253668 4934 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:43:09 crc kubenswrapper[4934]: E1227 07:43:09.253712 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 27 07:43:09 crc kubenswrapper[4934]: E1227 07:43:09.253742 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-27 07:43:41.253721557 +0000 UTC m=+82.074162181 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:43:09 crc kubenswrapper[4934]: E1227 07:43:09.253753 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 27 07:43:09 crc kubenswrapper[4934]: E1227 07:43:09.253774 4934 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:43:09 crc kubenswrapper[4934]: E1227 07:43:09.253844 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-27 07:43:41.25382011 +0000 UTC m=+82.074260744 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.353226 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.353266 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.353275 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.353291 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.353302 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:09Z","lastTransitionTime":"2025-12-27T07:43:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.456639 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.456700 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.456716 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.456738 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.456755 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:09Z","lastTransitionTime":"2025-12-27T07:43:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.467304 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.467341 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:09 crc kubenswrapper[4934]: E1227 07:43:09.467474 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.467680 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:09 crc kubenswrapper[4934]: E1227 07:43:09.467783 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:09 crc kubenswrapper[4934]: E1227 07:43:09.467985 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.487066 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:09Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.507145 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:09Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.522164 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:09Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.538391 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"304709d3-05ab-4d27-8722-f94053159a29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jjlqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:09Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.556723 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:09Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.559243 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.559281 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.559298 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.559319 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.559336 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:09Z","lastTransitionTime":"2025-12-27T07:43:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.570945 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:09Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.597408 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:04Z\\\",\\\"message\\\":\\\":43:04.648259 6555 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1227 07:43:04.648311 6555 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1227 07:43:04.648332 6555 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1227 07:43:04.648368 6555 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1227 07:43:04.648417 6555 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1227 07:43:04.648433 6555 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1227 07:43:04.648472 6555 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1227 07:43:04.648490 6555 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1227 07:43:04.648506 6555 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1227 07:43:04.648525 6555 handler.go:208] Removed *v1.Node event handler 2\\\\nI1227 07:43:04.648539 6555 handler.go:208] Removed *v1.Node event handler 7\\\\nI1227 07:43:04.648553 6555 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1227 07:43:04.648568 6555 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1227 07:43:04.649724 6555 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1227 07:43:04.649797 6555 factory.go:656] Stopping watch factory\\\\nI1227 07:43:04.649819 6555 ovnkube.go:599] Stopped ovnkube\\\\nI1227 07:43:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:43:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:09Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.612550 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd63c24f-fd03-4148-a9a0-a5b86882b52a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe31fde827f2497b12876938f30cc22d79ece0e3e2ab57a034266de29d28266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a5ed5db5f0c41fa30232392526be2a603c1fdc408aef12e8296deea70899805\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g2dnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:09Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.633660 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:09Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.652668 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:09Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.662326 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.662390 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.662407 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.662430 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.662447 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:09Z","lastTransitionTime":"2025-12-27T07:43:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.676977 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:09Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.696459 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:09Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.714755 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:09Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.729315 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:09Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.746616 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:09Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.765113 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.765166 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.765188 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.765220 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.765242 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:09Z","lastTransitionTime":"2025-12-27T07:43:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.767011 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:09Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.870144 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.870210 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.870230 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.870256 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.870273 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:09Z","lastTransitionTime":"2025-12-27T07:43:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.974120 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.974202 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.974220 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.974241 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:09 crc kubenswrapper[4934]: I1227 07:43:09.974297 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:09Z","lastTransitionTime":"2025-12-27T07:43:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.078006 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.078423 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.078441 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.078464 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.078482 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:10Z","lastTransitionTime":"2025-12-27T07:43:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.183209 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.183260 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.183286 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.183307 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.183321 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:10Z","lastTransitionTime":"2025-12-27T07:43:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.286907 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.286973 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.286989 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.287011 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.287029 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:10Z","lastTransitionTime":"2025-12-27T07:43:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.391295 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.391370 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.391388 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.391413 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.391431 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:10Z","lastTransitionTime":"2025-12-27T07:43:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.466847 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:10 crc kubenswrapper[4934]: E1227 07:43:10.467056 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.495281 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.495332 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.495350 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.495372 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.495393 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:10Z","lastTransitionTime":"2025-12-27T07:43:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.598334 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.598382 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.598395 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.598412 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.598424 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:10Z","lastTransitionTime":"2025-12-27T07:43:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.702363 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.702428 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.702445 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.702473 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.702490 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:10Z","lastTransitionTime":"2025-12-27T07:43:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.805428 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.805495 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.805518 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.805547 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.805570 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:10Z","lastTransitionTime":"2025-12-27T07:43:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.908819 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.908882 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.908899 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.908924 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:10 crc kubenswrapper[4934]: I1227 07:43:10.908944 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:10Z","lastTransitionTime":"2025-12-27T07:43:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.012409 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.012474 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.012499 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.012528 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.012550 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:11Z","lastTransitionTime":"2025-12-27T07:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.115614 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.115669 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.115677 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.115693 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.115703 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:11Z","lastTransitionTime":"2025-12-27T07:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.218392 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.218445 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.218461 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.218484 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.218502 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:11Z","lastTransitionTime":"2025-12-27T07:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.322170 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.322230 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.322253 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.322282 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.322303 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:11Z","lastTransitionTime":"2025-12-27T07:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.425422 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.426253 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.426400 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.426541 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.426674 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:11Z","lastTransitionTime":"2025-12-27T07:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.467476 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.467567 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.467507 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:11 crc kubenswrapper[4934]: E1227 07:43:11.467678 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:11 crc kubenswrapper[4934]: E1227 07:43:11.467798 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:11 crc kubenswrapper[4934]: E1227 07:43:11.467896 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.530387 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.530461 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.530484 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.530513 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.530536 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:11Z","lastTransitionTime":"2025-12-27T07:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.633561 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.633612 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.633629 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.633652 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.633669 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:11Z","lastTransitionTime":"2025-12-27T07:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.736804 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.736882 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.736909 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.736940 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.737005 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:11Z","lastTransitionTime":"2025-12-27T07:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.839554 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.839617 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.839639 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.839669 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.839694 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:11Z","lastTransitionTime":"2025-12-27T07:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.901179 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.901234 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.901251 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.901275 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.901293 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:11Z","lastTransitionTime":"2025-12-27T07:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:11 crc kubenswrapper[4934]: E1227 07:43:11.923349 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:11Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.929904 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.930163 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.930319 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.930452 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.930612 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:11Z","lastTransitionTime":"2025-12-27T07:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:11 crc kubenswrapper[4934]: E1227 07:43:11.951553 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:11Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.957372 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.957453 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.957479 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.957510 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.957535 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:11Z","lastTransitionTime":"2025-12-27T07:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:11 crc kubenswrapper[4934]: E1227 07:43:11.972330 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:11Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.977065 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.977162 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.977180 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.977204 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:11 crc kubenswrapper[4934]: I1227 07:43:11.977222 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:11Z","lastTransitionTime":"2025-12-27T07:43:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:12 crc kubenswrapper[4934]: E1227 07:43:12.007648 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:12Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.013002 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.013066 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.013117 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.013145 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.013168 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:12Z","lastTransitionTime":"2025-12-27T07:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:12 crc kubenswrapper[4934]: E1227 07:43:12.033587 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:12Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:12 crc kubenswrapper[4934]: E1227 07:43:12.033906 4934 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.036487 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.036575 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.036600 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.036631 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.036655 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:12Z","lastTransitionTime":"2025-12-27T07:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.139921 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.139983 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.140002 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.140028 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.140051 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:12Z","lastTransitionTime":"2025-12-27T07:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.242965 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.243025 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.243042 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.243066 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.243109 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:12Z","lastTransitionTime":"2025-12-27T07:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.345578 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.345692 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.345710 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.345751 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.345770 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:12Z","lastTransitionTime":"2025-12-27T07:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.448747 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.448798 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.448811 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.448829 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.448843 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:12Z","lastTransitionTime":"2025-12-27T07:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.467446 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:12 crc kubenswrapper[4934]: E1227 07:43:12.467653 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.551157 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.551213 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.551233 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.551256 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.551274 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:12Z","lastTransitionTime":"2025-12-27T07:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.654266 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.654324 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.654350 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.654381 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.654404 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:12Z","lastTransitionTime":"2025-12-27T07:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.757157 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.757218 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.757238 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.757264 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.757287 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:12Z","lastTransitionTime":"2025-12-27T07:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.859965 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.860037 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.860055 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.860110 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.860130 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:12Z","lastTransitionTime":"2025-12-27T07:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.962857 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.962935 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.962959 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.962989 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:12 crc kubenswrapper[4934]: I1227 07:43:12.963014 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:12Z","lastTransitionTime":"2025-12-27T07:43:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.065788 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.065844 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.065861 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.065883 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.065903 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:13Z","lastTransitionTime":"2025-12-27T07:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.169048 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.169136 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.169151 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.169169 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.169184 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:13Z","lastTransitionTime":"2025-12-27T07:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.272131 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.272187 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.272204 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.272227 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.272244 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:13Z","lastTransitionTime":"2025-12-27T07:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.374249 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.374319 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.374335 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.374359 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.374376 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:13Z","lastTransitionTime":"2025-12-27T07:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.467051 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.467058 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:13 crc kubenswrapper[4934]: E1227 07:43:13.467322 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.467113 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:13 crc kubenswrapper[4934]: E1227 07:43:13.467459 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:13 crc kubenswrapper[4934]: E1227 07:43:13.467588 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.477032 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.477109 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.477131 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.477154 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.477170 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:13Z","lastTransitionTime":"2025-12-27T07:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.580558 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.580619 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.580636 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.580661 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.580681 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:13Z","lastTransitionTime":"2025-12-27T07:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.684199 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.684272 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.684296 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.684326 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.684358 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:13Z","lastTransitionTime":"2025-12-27T07:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.787073 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.787165 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.787188 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.787213 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.787230 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:13Z","lastTransitionTime":"2025-12-27T07:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.890323 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.890381 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.890400 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.890424 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.890444 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:13Z","lastTransitionTime":"2025-12-27T07:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.993500 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.993562 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.993591 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.993622 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:13 crc kubenswrapper[4934]: I1227 07:43:13.993644 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:13Z","lastTransitionTime":"2025-12-27T07:43:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.096564 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.096627 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.096644 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.096673 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.096692 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:14Z","lastTransitionTime":"2025-12-27T07:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.150517 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.167334 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.176478 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:14Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.197373 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:14Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.199345 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.199408 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.199425 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.199454 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.199472 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:14Z","lastTransitionTime":"2025-12-27T07:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.218963 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:14Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.237061 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:14Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.254569 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:14Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.268729 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"304709d3-05ab-4d27-8722-f94053159a29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jjlqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:14Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.286402 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:14Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.302879 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:14Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.303045 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.303415 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.303442 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.303472 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.303495 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:14Z","lastTransitionTime":"2025-12-27T07:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.316762 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:14Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.333259 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd63c24f-fd03-4148-a9a0-a5b86882b52a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe31fde827f2497b12876938f30cc22d79ece0e3e2ab57a034266de29d28266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a5ed5db5f0c41fa30232392526be2a603c1fdc408aef12e8296deea70899805\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g2dnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:14Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.353629 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:14Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.372059 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:14Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.403073 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:04Z\\\",\\\"message\\\":\\\":43:04.648259 6555 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1227 07:43:04.648311 6555 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1227 07:43:04.648332 6555 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1227 07:43:04.648368 6555 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1227 07:43:04.648417 6555 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1227 07:43:04.648433 6555 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1227 07:43:04.648472 6555 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1227 07:43:04.648490 6555 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1227 07:43:04.648506 6555 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1227 07:43:04.648525 6555 handler.go:208] Removed *v1.Node event handler 2\\\\nI1227 07:43:04.648539 6555 handler.go:208] Removed *v1.Node event handler 7\\\\nI1227 07:43:04.648553 6555 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1227 07:43:04.648568 6555 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1227 07:43:04.649724 6555 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1227 07:43:04.649797 6555 factory.go:656] Stopping watch factory\\\\nI1227 07:43:04.649819 6555 ovnkube.go:599] Stopped ovnkube\\\\nI1227 07:43:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:43:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:14Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.405666 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.405802 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.405904 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.406134 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.406276 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:14Z","lastTransitionTime":"2025-12-27T07:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.426062 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:14Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.444906 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:14Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.467258 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:14 crc kubenswrapper[4934]: E1227 07:43:14.467493 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.469605 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:14Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.510141 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.510196 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.510213 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.510237 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.510254 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:14Z","lastTransitionTime":"2025-12-27T07:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.613028 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.613458 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.613675 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.613907 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.614149 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:14Z","lastTransitionTime":"2025-12-27T07:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.717597 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.717657 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.717679 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.717711 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.717733 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:14Z","lastTransitionTime":"2025-12-27T07:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.820606 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.820675 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.820694 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.820719 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.820741 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:14Z","lastTransitionTime":"2025-12-27T07:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.923592 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.923661 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.923685 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.923723 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:14 crc kubenswrapper[4934]: I1227 07:43:14.923748 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:14Z","lastTransitionTime":"2025-12-27T07:43:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.027147 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.027202 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.027223 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.027255 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.027281 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:15Z","lastTransitionTime":"2025-12-27T07:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.130643 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.130707 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.130741 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.130781 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.130805 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:15Z","lastTransitionTime":"2025-12-27T07:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.233547 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.233618 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.233641 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.233671 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.233694 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:15Z","lastTransitionTime":"2025-12-27T07:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.336985 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.337027 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.337112 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.337143 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.337155 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:15Z","lastTransitionTime":"2025-12-27T07:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.440182 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.440242 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.440264 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.440292 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.440311 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:15Z","lastTransitionTime":"2025-12-27T07:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.467273 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.467422 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:15 crc kubenswrapper[4934]: E1227 07:43:15.467473 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:15 crc kubenswrapper[4934]: E1227 07:43:15.467623 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.467730 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:15 crc kubenswrapper[4934]: E1227 07:43:15.467848 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.542766 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.542833 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.542858 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.542885 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.542903 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:15Z","lastTransitionTime":"2025-12-27T07:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.645550 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.645622 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.645649 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.645679 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.645702 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:15Z","lastTransitionTime":"2025-12-27T07:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.748953 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.749023 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.749046 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.749074 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.749123 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:15Z","lastTransitionTime":"2025-12-27T07:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.851801 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.851845 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.851857 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.851873 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.851885 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:15Z","lastTransitionTime":"2025-12-27T07:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.954472 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.954529 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.954550 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.954574 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:15 crc kubenswrapper[4934]: I1227 07:43:15.954593 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:15Z","lastTransitionTime":"2025-12-27T07:43:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.056832 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.056903 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.056921 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.056944 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.056963 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:16Z","lastTransitionTime":"2025-12-27T07:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.159801 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.159870 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.159894 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.159924 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.159945 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:16Z","lastTransitionTime":"2025-12-27T07:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.262220 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.262261 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.262273 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.262288 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.262300 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:16Z","lastTransitionTime":"2025-12-27T07:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.365330 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.365391 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.365410 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.365432 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.365451 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:16Z","lastTransitionTime":"2025-12-27T07:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.467020 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:16 crc kubenswrapper[4934]: E1227 07:43:16.467290 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.468833 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.468886 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.468908 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.468934 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.468955 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:16Z","lastTransitionTime":"2025-12-27T07:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.571718 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.571805 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.571832 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.571862 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.571884 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:16Z","lastTransitionTime":"2025-12-27T07:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.675029 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.675117 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.675134 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.675158 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.675173 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:16Z","lastTransitionTime":"2025-12-27T07:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.777526 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.777598 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.777622 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.777652 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.777675 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:16Z","lastTransitionTime":"2025-12-27T07:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.880603 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.880640 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.880652 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.880667 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.880679 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:16Z","lastTransitionTime":"2025-12-27T07:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.983511 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.983561 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.983576 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.983599 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:16 crc kubenswrapper[4934]: I1227 07:43:16.983617 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:16Z","lastTransitionTime":"2025-12-27T07:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.086634 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.086670 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.086681 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.086703 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.086716 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:17Z","lastTransitionTime":"2025-12-27T07:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.189655 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.189702 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.189714 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.189733 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.189747 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:17Z","lastTransitionTime":"2025-12-27T07:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.292990 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.293042 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.293058 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.293079 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.293128 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:17Z","lastTransitionTime":"2025-12-27T07:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.395565 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.395645 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.395662 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.395688 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.395706 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:17Z","lastTransitionTime":"2025-12-27T07:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.466648 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.466651 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.466858 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:17 crc kubenswrapper[4934]: E1227 07:43:17.466759 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:17 crc kubenswrapper[4934]: E1227 07:43:17.467024 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:17 crc kubenswrapper[4934]: E1227 07:43:17.467257 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.497824 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.497883 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.497896 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.497911 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.497921 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:17Z","lastTransitionTime":"2025-12-27T07:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.600201 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.600262 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.600273 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.600309 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.600323 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:17Z","lastTransitionTime":"2025-12-27T07:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.702821 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.702880 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.702893 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.702908 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.702919 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:17Z","lastTransitionTime":"2025-12-27T07:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.805753 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.805798 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.805811 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.805829 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.805841 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:17Z","lastTransitionTime":"2025-12-27T07:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.907652 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.907727 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.907741 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.907760 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:17 crc kubenswrapper[4934]: I1227 07:43:17.907772 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:17Z","lastTransitionTime":"2025-12-27T07:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.010525 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.010566 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.010577 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.010596 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.010607 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:18Z","lastTransitionTime":"2025-12-27T07:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.113777 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.113826 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.113838 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.113857 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.113869 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:18Z","lastTransitionTime":"2025-12-27T07:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.216884 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.216938 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.216957 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.216982 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.216999 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:18Z","lastTransitionTime":"2025-12-27T07:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.319837 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.320141 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.320240 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.320340 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.320430 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:18Z","lastTransitionTime":"2025-12-27T07:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.423544 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.423611 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.423629 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.423653 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.423670 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:18Z","lastTransitionTime":"2025-12-27T07:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.466501 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:18 crc kubenswrapper[4934]: E1227 07:43:18.467194 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.527240 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.527807 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.527933 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.528028 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.528141 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:18Z","lastTransitionTime":"2025-12-27T07:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.631220 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.631281 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.631297 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.631321 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.631338 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:18Z","lastTransitionTime":"2025-12-27T07:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.734365 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.734441 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.734458 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.734482 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.734515 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:18Z","lastTransitionTime":"2025-12-27T07:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.837859 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.837913 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.837927 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.837947 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.837961 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:18Z","lastTransitionTime":"2025-12-27T07:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.941613 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.941660 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.941675 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.941695 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:18 crc kubenswrapper[4934]: I1227 07:43:18.941708 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:18Z","lastTransitionTime":"2025-12-27T07:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.044628 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.044688 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.044705 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.044762 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.044781 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:19Z","lastTransitionTime":"2025-12-27T07:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.147484 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.148553 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.148980 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.149270 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.149513 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:19Z","lastTransitionTime":"2025-12-27T07:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.252148 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.252397 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.252535 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.252652 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.252755 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:19Z","lastTransitionTime":"2025-12-27T07:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.355320 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.355652 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.355922 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.356065 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.356245 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:19Z","lastTransitionTime":"2025-12-27T07:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.459448 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.459507 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.459524 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.459547 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.459563 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:19Z","lastTransitionTime":"2025-12-27T07:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.467346 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.467416 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:19 crc kubenswrapper[4934]: E1227 07:43:19.467486 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:19 crc kubenswrapper[4934]: E1227 07:43:19.467615 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.468392 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:19 crc kubenswrapper[4934]: E1227 07:43:19.468564 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.488895 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ee2490b-5c06-4f10-a362-b5fcb71f32a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80370fd8607ddb05cccbe7383b94c5a4c1a874299bbbc7f665805ae5edf73ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e461c87813b05a23e571f459ad229874b607be5916ab31f502d476655d5f3d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://243b20c55923c82f8d7a735a9bd0656bfe13008645b5b637c9d0f3890b5890d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:19Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.509185 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:19Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.527839 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:19Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.544129 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:19Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.560923 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"304709d3-05ab-4d27-8722-f94053159a29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jjlqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:19Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.565741 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.565805 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.565832 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.565881 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.565909 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:19Z","lastTransitionTime":"2025-12-27T07:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.585911 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:19Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.602619 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:19Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.630740 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:04Z\\\",\\\"message\\\":\\\":43:04.648259 6555 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1227 07:43:04.648311 6555 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1227 07:43:04.648332 6555 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1227 07:43:04.648368 6555 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1227 07:43:04.648417 6555 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1227 07:43:04.648433 6555 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1227 07:43:04.648472 6555 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1227 07:43:04.648490 6555 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1227 07:43:04.648506 6555 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1227 07:43:04.648525 6555 handler.go:208] Removed *v1.Node event handler 2\\\\nI1227 07:43:04.648539 6555 handler.go:208] Removed *v1.Node event handler 7\\\\nI1227 07:43:04.648553 6555 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1227 07:43:04.648568 6555 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1227 07:43:04.649724 6555 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1227 07:43:04.649797 6555 factory.go:656] Stopping watch factory\\\\nI1227 07:43:04.649819 6555 ovnkube.go:599] Stopped ovnkube\\\\nI1227 07:43:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:43:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:19Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.641978 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd63c24f-fd03-4148-a9a0-a5b86882b52a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe31fde827f2497b12876938f30cc22d79ece0e3e2ab57a034266de29d28266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a5ed5db5f0c41fa30232392526be2a603c1fdc408aef12e8296deea70899805\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g2dnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:19Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.654554 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:19Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.669061 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.669175 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.669202 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.669233 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.669256 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:19Z","lastTransitionTime":"2025-12-27T07:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.670679 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:19Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.686008 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:19Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.701221 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:19Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.719978 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:19Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.731052 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:19Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.745436 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:19Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.760195 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:19Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.771973 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.772015 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.772023 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.772035 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.772043 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:19Z","lastTransitionTime":"2025-12-27T07:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.874623 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.874649 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.874657 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.874669 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.874677 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:19Z","lastTransitionTime":"2025-12-27T07:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.978272 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.978359 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.978380 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.978460 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:19 crc kubenswrapper[4934]: I1227 07:43:19.978482 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:19Z","lastTransitionTime":"2025-12-27T07:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.081718 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.081780 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.081804 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.081834 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.081857 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:20Z","lastTransitionTime":"2025-12-27T07:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.184463 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.184525 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.184542 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.184564 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.184581 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:20Z","lastTransitionTime":"2025-12-27T07:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.286714 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.286784 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.286801 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.286826 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.286847 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:20Z","lastTransitionTime":"2025-12-27T07:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.388952 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.388991 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.389000 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.389013 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.389021 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:20Z","lastTransitionTime":"2025-12-27T07:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.466360 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:20 crc kubenswrapper[4934]: E1227 07:43:20.466536 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.492131 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.492171 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.492186 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.492203 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.492216 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:20Z","lastTransitionTime":"2025-12-27T07:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.596799 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.597246 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.597266 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.597288 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.597304 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:20Z","lastTransitionTime":"2025-12-27T07:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.699830 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.699897 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.699914 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.699939 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.699956 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:20Z","lastTransitionTime":"2025-12-27T07:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.803033 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.803131 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.803155 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.803190 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.803214 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:20Z","lastTransitionTime":"2025-12-27T07:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.905560 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.905617 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.905632 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.905651 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:20 crc kubenswrapper[4934]: I1227 07:43:20.905665 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:20Z","lastTransitionTime":"2025-12-27T07:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.008873 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.008925 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.008937 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.008954 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.008967 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:21Z","lastTransitionTime":"2025-12-27T07:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.111396 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.111443 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.111456 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.111474 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.111487 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:21Z","lastTransitionTime":"2025-12-27T07:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.214177 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.214229 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.214244 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.214268 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.214284 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:21Z","lastTransitionTime":"2025-12-27T07:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.319021 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.319075 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.319113 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.319134 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.319146 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:21Z","lastTransitionTime":"2025-12-27T07:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.421366 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.421401 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.421412 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.421427 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.421438 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:21Z","lastTransitionTime":"2025-12-27T07:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.468375 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.468463 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:21 crc kubenswrapper[4934]: E1227 07:43:21.468610 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.468651 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:21 crc kubenswrapper[4934]: E1227 07:43:21.468730 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:21 crc kubenswrapper[4934]: E1227 07:43:21.468792 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.523872 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.523912 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.523926 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.523941 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.523951 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:21Z","lastTransitionTime":"2025-12-27T07:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.627036 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.627126 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.627145 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.627166 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.627183 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:21Z","lastTransitionTime":"2025-12-27T07:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.729778 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.729820 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.729835 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.729853 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.729869 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:21Z","lastTransitionTime":"2025-12-27T07:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.831531 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.831561 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.831568 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.831580 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.831588 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:21Z","lastTransitionTime":"2025-12-27T07:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.933583 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.933610 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.933617 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.933629 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:21 crc kubenswrapper[4934]: I1227 07:43:21.933639 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:21Z","lastTransitionTime":"2025-12-27T07:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.035808 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.035848 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.035860 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.035874 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.035886 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:22Z","lastTransitionTime":"2025-12-27T07:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.138194 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.138333 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.138353 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.138379 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.138397 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:22Z","lastTransitionTime":"2025-12-27T07:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.240619 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.240677 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.240700 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.240721 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.240735 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:22Z","lastTransitionTime":"2025-12-27T07:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.243880 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.243935 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.243953 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.243973 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.243989 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:22Z","lastTransitionTime":"2025-12-27T07:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:22 crc kubenswrapper[4934]: E1227 07:43:22.257539 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:22Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.261052 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.261122 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.261139 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.261161 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.261177 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:22Z","lastTransitionTime":"2025-12-27T07:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:22 crc kubenswrapper[4934]: E1227 07:43:22.275684 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:22Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.278529 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.278554 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.278563 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.278575 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.278586 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:22Z","lastTransitionTime":"2025-12-27T07:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:22 crc kubenswrapper[4934]: E1227 07:43:22.294127 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:22Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.297707 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.297751 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.297763 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.297785 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.297798 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:22Z","lastTransitionTime":"2025-12-27T07:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:22 crc kubenswrapper[4934]: E1227 07:43:22.311381 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:22Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.315451 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.315491 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.315507 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.315527 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.315543 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:22Z","lastTransitionTime":"2025-12-27T07:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:22 crc kubenswrapper[4934]: E1227 07:43:22.329075 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:22Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:22 crc kubenswrapper[4934]: E1227 07:43:22.329199 4934 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.343804 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.343838 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.343850 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.343872 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.343883 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:22Z","lastTransitionTime":"2025-12-27T07:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.446591 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.446636 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.446647 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.446664 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.446675 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:22Z","lastTransitionTime":"2025-12-27T07:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.467122 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:22 crc kubenswrapper[4934]: E1227 07:43:22.467317 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.549188 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.549225 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.549237 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.549253 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.549267 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:22Z","lastTransitionTime":"2025-12-27T07:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.651210 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.651254 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.651266 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.651281 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.651294 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:22Z","lastTransitionTime":"2025-12-27T07:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.753554 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.753591 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.753601 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.753616 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.753627 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:22Z","lastTransitionTime":"2025-12-27T07:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.855518 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.855588 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.855606 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.855628 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.855645 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:22Z","lastTransitionTime":"2025-12-27T07:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.958040 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.958129 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.958152 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.958183 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:22 crc kubenswrapper[4934]: I1227 07:43:22.958205 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:22Z","lastTransitionTime":"2025-12-27T07:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.060954 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.060998 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.061010 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.061034 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.061049 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:23Z","lastTransitionTime":"2025-12-27T07:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.163921 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.163979 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.163995 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.164021 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.164039 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:23Z","lastTransitionTime":"2025-12-27T07:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.265865 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.265912 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.265925 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.265943 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.265957 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:23Z","lastTransitionTime":"2025-12-27T07:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.369043 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.369115 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.369132 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.369155 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.369167 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:23Z","lastTransitionTime":"2025-12-27T07:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.466953 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.467011 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.466961 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:23 crc kubenswrapper[4934]: E1227 07:43:23.467111 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:23 crc kubenswrapper[4934]: E1227 07:43:23.467178 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:23 crc kubenswrapper[4934]: E1227 07:43:23.467589 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.467823 4934 scope.go:117] "RemoveContainer" containerID="8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05" Dec 27 07:43:23 crc kubenswrapper[4934]: E1227 07:43:23.468059 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.470676 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.470699 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.470712 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.470727 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.470737 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:23Z","lastTransitionTime":"2025-12-27T07:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.573000 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.573038 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.573048 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.573064 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.573075 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:23Z","lastTransitionTime":"2025-12-27T07:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.675659 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.675704 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.675717 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.675736 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.675750 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:23Z","lastTransitionTime":"2025-12-27T07:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.778140 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.778515 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.778647 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.778781 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.778905 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:23Z","lastTransitionTime":"2025-12-27T07:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.881287 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.881331 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.881343 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.881362 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.881375 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:23Z","lastTransitionTime":"2025-12-27T07:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.983782 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.983855 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.983866 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.983885 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:23 crc kubenswrapper[4934]: I1227 07:43:23.983896 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:23Z","lastTransitionTime":"2025-12-27T07:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.086365 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.086438 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.086460 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.086490 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.086512 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:24Z","lastTransitionTime":"2025-12-27T07:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.188636 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.188687 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.188703 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.188728 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.188744 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:24Z","lastTransitionTime":"2025-12-27T07:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.290728 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.290799 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.290810 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.290824 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.290835 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:24Z","lastTransitionTime":"2025-12-27T07:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.393680 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.393740 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.393756 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.393779 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.393852 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:24Z","lastTransitionTime":"2025-12-27T07:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.466812 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:24 crc kubenswrapper[4934]: E1227 07:43:24.467047 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.496719 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.496767 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.496783 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.496800 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.496812 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:24Z","lastTransitionTime":"2025-12-27T07:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.598899 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.598959 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.598976 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.599000 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.599018 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:24Z","lastTransitionTime":"2025-12-27T07:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.625649 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs\") pod \"network-metrics-daemon-jjlqg\" (UID: \"304709d3-05ab-4d27-8722-f94053159a29\") " pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:24 crc kubenswrapper[4934]: E1227 07:43:24.625904 4934 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 27 07:43:24 crc kubenswrapper[4934]: E1227 07:43:24.625981 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs podName:304709d3-05ab-4d27-8722-f94053159a29 nodeName:}" failed. No retries permitted until 2025-12-27 07:43:56.625958199 +0000 UTC m=+97.446398833 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs") pod "network-metrics-daemon-jjlqg" (UID: "304709d3-05ab-4d27-8722-f94053159a29") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.701145 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.701194 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.701209 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.701231 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.701248 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:24Z","lastTransitionTime":"2025-12-27T07:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.803257 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.803391 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.803402 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.803416 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.803425 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:24Z","lastTransitionTime":"2025-12-27T07:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.905267 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.905292 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.905299 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.905309 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:24 crc kubenswrapper[4934]: I1227 07:43:24.905318 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:24Z","lastTransitionTime":"2025-12-27T07:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.007500 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.007534 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.007541 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.007553 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.007562 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:25Z","lastTransitionTime":"2025-12-27T07:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.109855 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.109917 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.109934 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.109956 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.109973 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:25Z","lastTransitionTime":"2025-12-27T07:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.212357 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.212400 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.212411 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.212430 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.212444 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:25Z","lastTransitionTime":"2025-12-27T07:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.314928 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.314972 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.314982 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.314996 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.315005 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:25Z","lastTransitionTime":"2025-12-27T07:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.417526 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.417573 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.417587 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.417605 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.417617 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:25Z","lastTransitionTime":"2025-12-27T07:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.466759 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.466799 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.466830 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:25 crc kubenswrapper[4934]: E1227 07:43:25.466864 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:25 crc kubenswrapper[4934]: E1227 07:43:25.466955 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:25 crc kubenswrapper[4934]: E1227 07:43:25.467095 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.519768 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.519804 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.519813 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.519826 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.519837 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:25Z","lastTransitionTime":"2025-12-27T07:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.621968 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.622203 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.622278 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.622343 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.622402 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:25Z","lastTransitionTime":"2025-12-27T07:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.724140 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.724165 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.724173 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.724186 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.724196 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:25Z","lastTransitionTime":"2025-12-27T07:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.827303 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.827377 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.827396 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.827419 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.827437 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:25Z","lastTransitionTime":"2025-12-27T07:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.929232 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jzv2s_1bdd4b78-f615-4771-9448-8faa03dbc4f5/kube-multus/0.log" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.929339 4934 generic.go:334] "Generic (PLEG): container finished" podID="1bdd4b78-f615-4771-9448-8faa03dbc4f5" containerID="a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44" exitCode=1 Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.929396 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jzv2s" event={"ID":"1bdd4b78-f615-4771-9448-8faa03dbc4f5","Type":"ContainerDied","Data":"a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44"} Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.930209 4934 scope.go:117] "RemoveContainer" containerID="a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.930661 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.930717 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.930736 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.930757 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.930774 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:25Z","lastTransitionTime":"2025-12-27T07:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.945159 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"304709d3-05ab-4d27-8722-f94053159a29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jjlqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:25Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.964528 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ee2490b-5c06-4f10-a362-b5fcb71f32a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80370fd8607ddb05cccbe7383b94c5a4c1a874299bbbc7f665805ae5edf73ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e461c87813b05a23e571f459ad229874b607be5916ab31f502d476655d5f3d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://243b20c55923c82f8d7a735a9bd0656bfe13008645b5b637c9d0f3890b5890d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:25Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.980642 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:25Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:25 crc kubenswrapper[4934]: I1227 07:43:25.999119 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:25Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.011998 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:26Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.029998 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd63c24f-fd03-4148-a9a0-a5b86882b52a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe31fde827f2497b12876938f30cc22d79ece0e3e2ab57a034266de29d28266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a5ed5db5f0c41fa30232392526be2a603c1fdc408aef12e8296deea70899805\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g2dnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:26Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.035045 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.035100 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.035111 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.035125 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.035150 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:26Z","lastTransitionTime":"2025-12-27T07:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.058807 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:26Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.072511 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:26Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.089158 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:04Z\\\",\\\"message\\\":\\\":43:04.648259 6555 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1227 07:43:04.648311 6555 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1227 07:43:04.648332 6555 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1227 07:43:04.648368 6555 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1227 07:43:04.648417 6555 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1227 07:43:04.648433 6555 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1227 07:43:04.648472 6555 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1227 07:43:04.648490 6555 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1227 07:43:04.648506 6555 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1227 07:43:04.648525 6555 handler.go:208] Removed *v1.Node event handler 2\\\\nI1227 07:43:04.648539 6555 handler.go:208] Removed *v1.Node event handler 7\\\\nI1227 07:43:04.648553 6555 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1227 07:43:04.648568 6555 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1227 07:43:04.649724 6555 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1227 07:43:04.649797 6555 factory.go:656] Stopping watch factory\\\\nI1227 07:43:04.649819 6555 ovnkube.go:599] Stopped ovnkube\\\\nI1227 07:43:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:43:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:26Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.109936 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:26Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.127222 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:26Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.137470 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.137524 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.137539 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.137563 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.137579 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:26Z","lastTransitionTime":"2025-12-27T07:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.140922 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:26Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.152815 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:25Z\\\",\\\"message\\\":\\\"2025-12-27T07:42:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b83db5b0-0703-4f20-99c4-88b5d3e7ba75\\\\n2025-12-27T07:42:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b83db5b0-0703-4f20-99c4-88b5d3e7ba75 to /host/opt/cni/bin/\\\\n2025-12-27T07:42:40Z [verbose] multus-daemon started\\\\n2025-12-27T07:42:40Z [verbose] Readiness Indicator file check\\\\n2025-12-27T07:43:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:26Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.163977 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:26Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.174397 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:26Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.190803 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:26Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.201118 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:26Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.239426 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.239457 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.239466 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.239479 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.239488 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:26Z","lastTransitionTime":"2025-12-27T07:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.341981 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.342022 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.342035 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.342051 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.342062 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:26Z","lastTransitionTime":"2025-12-27T07:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.445149 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.445192 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.445203 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.445219 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.445231 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:26Z","lastTransitionTime":"2025-12-27T07:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.466981 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:26 crc kubenswrapper[4934]: E1227 07:43:26.467152 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.547374 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.547409 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.547420 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.547435 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.547449 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:26Z","lastTransitionTime":"2025-12-27T07:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.650045 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.650082 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.650109 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.650125 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.650137 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:26Z","lastTransitionTime":"2025-12-27T07:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.752176 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.752247 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.752270 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.752300 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.752324 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:26Z","lastTransitionTime":"2025-12-27T07:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.855341 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.855721 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.855743 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.855767 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.855788 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:26Z","lastTransitionTime":"2025-12-27T07:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.934786 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jzv2s_1bdd4b78-f615-4771-9448-8faa03dbc4f5/kube-multus/0.log" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.934846 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jzv2s" event={"ID":"1bdd4b78-f615-4771-9448-8faa03dbc4f5","Type":"ContainerStarted","Data":"dd93f8a550f1648117195c4caac3c1e150c84a1c2dba90ae1ba1a4ecc1bd284b"} Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.951892 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:26Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.958128 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.958158 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.958166 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.958182 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.958193 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:26Z","lastTransitionTime":"2025-12-27T07:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.964910 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:26Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.981582 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:04Z\\\",\\\"message\\\":\\\":43:04.648259 6555 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1227 07:43:04.648311 6555 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1227 07:43:04.648332 6555 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1227 07:43:04.648368 6555 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1227 07:43:04.648417 6555 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1227 07:43:04.648433 6555 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1227 07:43:04.648472 6555 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1227 07:43:04.648490 6555 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1227 07:43:04.648506 6555 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1227 07:43:04.648525 6555 handler.go:208] Removed *v1.Node event handler 2\\\\nI1227 07:43:04.648539 6555 handler.go:208] Removed *v1.Node event handler 7\\\\nI1227 07:43:04.648553 6555 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1227 07:43:04.648568 6555 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1227 07:43:04.649724 6555 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1227 07:43:04.649797 6555 factory.go:656] Stopping watch factory\\\\nI1227 07:43:04.649819 6555 ovnkube.go:599] Stopped ovnkube\\\\nI1227 07:43:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:43:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:26Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:26 crc kubenswrapper[4934]: I1227 07:43:26.995485 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd63c24f-fd03-4148-a9a0-a5b86882b52a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe31fde827f2497b12876938f30cc22d79ece0e3e2ab57a034266de29d28266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a5ed5db5f0c41fa30232392526be2a603c1fdc408aef12e8296deea70899805\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g2dnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:26Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.012805 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:27Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.023548 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:27Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.035525 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:27Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.047604 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:27Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.059815 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.059813 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:27Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.059894 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.059909 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.059925 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.059937 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:27Z","lastTransitionTime":"2025-12-27T07:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.068572 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:27Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.077670 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:27Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.088120 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd93f8a550f1648117195c4caac3c1e150c84a1c2dba90ae1ba1a4ecc1bd284b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:25Z\\\",\\\"message\\\":\\\"2025-12-27T07:42:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b83db5b0-0703-4f20-99c4-88b5d3e7ba75\\\\n2025-12-27T07:42:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b83db5b0-0703-4f20-99c4-88b5d3e7ba75 to /host/opt/cni/bin/\\\\n2025-12-27T07:42:40Z [verbose] multus-daemon started\\\\n2025-12-27T07:42:40Z [verbose] Readiness Indicator file check\\\\n2025-12-27T07:43:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:43:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:27Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.097228 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ee2490b-5c06-4f10-a362-b5fcb71f32a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80370fd8607ddb05cccbe7383b94c5a4c1a874299bbbc7f665805ae5edf73ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e461c87813b05a23e571f459ad229874b607be5916ab31f502d476655d5f3d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://243b20c55923c82f8d7a735a9bd0656bfe13008645b5b637c9d0f3890b5890d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:27Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.107643 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:27Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.119145 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:27Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.127831 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:27Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.135852 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"304709d3-05ab-4d27-8722-f94053159a29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jjlqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:27Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.161900 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.161935 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.161946 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.161961 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.161972 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:27Z","lastTransitionTime":"2025-12-27T07:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.264243 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.264279 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.264289 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.264303 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.264314 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:27Z","lastTransitionTime":"2025-12-27T07:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.366487 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.366529 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.366540 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.366553 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.366563 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:27Z","lastTransitionTime":"2025-12-27T07:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.466783 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:27 crc kubenswrapper[4934]: E1227 07:43:27.466878 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.467041 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.467129 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:27 crc kubenswrapper[4934]: E1227 07:43:27.467397 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:27 crc kubenswrapper[4934]: E1227 07:43:27.467639 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.468137 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.468154 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.468161 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.468170 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.468179 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:27Z","lastTransitionTime":"2025-12-27T07:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.569942 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.569977 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.569987 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.570000 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.570011 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:27Z","lastTransitionTime":"2025-12-27T07:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.673153 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.673209 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.673224 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.673248 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.673264 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:27Z","lastTransitionTime":"2025-12-27T07:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.776471 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.776545 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.776563 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.776587 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.776604 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:27Z","lastTransitionTime":"2025-12-27T07:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.878705 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.878771 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.878794 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.878821 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.878838 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:27Z","lastTransitionTime":"2025-12-27T07:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.980837 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.980885 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.980906 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.980927 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:27 crc kubenswrapper[4934]: I1227 07:43:27.980939 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:27Z","lastTransitionTime":"2025-12-27T07:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.083412 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.083462 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.083475 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.083493 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.083505 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:28Z","lastTransitionTime":"2025-12-27T07:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.186562 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.186615 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.186630 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.186651 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.186669 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:28Z","lastTransitionTime":"2025-12-27T07:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.289553 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.289601 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.289618 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.289643 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.289660 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:28Z","lastTransitionTime":"2025-12-27T07:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.391826 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.391905 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.391924 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.391948 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.391966 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:28Z","lastTransitionTime":"2025-12-27T07:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.466980 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:28 crc kubenswrapper[4934]: E1227 07:43:28.467141 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.493527 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.493585 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.493602 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.493624 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.493641 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:28Z","lastTransitionTime":"2025-12-27T07:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.596127 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.596188 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.596206 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.596230 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.596248 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:28Z","lastTransitionTime":"2025-12-27T07:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.698789 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.698860 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.698895 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.698923 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.698944 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:28Z","lastTransitionTime":"2025-12-27T07:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.801784 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.801823 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.801833 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.801853 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.801871 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:28Z","lastTransitionTime":"2025-12-27T07:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.904338 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.904395 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.904412 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.904478 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:28 crc kubenswrapper[4934]: I1227 07:43:28.904498 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:28Z","lastTransitionTime":"2025-12-27T07:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.007323 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.007381 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.007398 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.007423 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.007442 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:29Z","lastTransitionTime":"2025-12-27T07:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.110214 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.110277 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.110293 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.110317 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.110334 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:29Z","lastTransitionTime":"2025-12-27T07:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.213528 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.213594 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.213611 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.213635 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.213652 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:29Z","lastTransitionTime":"2025-12-27T07:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.316685 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.316729 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.316745 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.316766 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.316781 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:29Z","lastTransitionTime":"2025-12-27T07:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.418731 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.418768 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.418780 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.418795 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.418808 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:29Z","lastTransitionTime":"2025-12-27T07:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.466376 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:29 crc kubenswrapper[4934]: E1227 07:43:29.466488 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.466539 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:29 crc kubenswrapper[4934]: E1227 07:43:29.466680 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.466745 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:29 crc kubenswrapper[4934]: E1227 07:43:29.466824 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.476639 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"304709d3-05ab-4d27-8722-f94053159a29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jjlqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:29Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.487640 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ee2490b-5c06-4f10-a362-b5fcb71f32a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80370fd8607ddb05cccbe7383b94c5a4c1a874299bbbc7f665805ae5edf73ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e461c87813b05a23e571f459ad229874b607be5916ab31f502d476655d5f3d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://243b20c55923c82f8d7a735a9bd0656bfe13008645b5b637c9d0f3890b5890d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:29Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.503325 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:29Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.519470 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:29Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.520594 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.520634 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.520652 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.520674 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.520690 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:29Z","lastTransitionTime":"2025-12-27T07:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.530941 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:29Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.543377 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd63c24f-fd03-4148-a9a0-a5b86882b52a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe31fde827f2497b12876938f30cc22d79ece0e3e2ab57a034266de29d28266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a5ed5db5f0c41fa30232392526be2a603c1fdc408aef12e8296deea70899805\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g2dnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:29Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.560134 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:29Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.578731 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:29Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.609585 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:04Z\\\",\\\"message\\\":\\\":43:04.648259 6555 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1227 07:43:04.648311 6555 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1227 07:43:04.648332 6555 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1227 07:43:04.648368 6555 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1227 07:43:04.648417 6555 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1227 07:43:04.648433 6555 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1227 07:43:04.648472 6555 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1227 07:43:04.648490 6555 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1227 07:43:04.648506 6555 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1227 07:43:04.648525 6555 handler.go:208] Removed *v1.Node event handler 2\\\\nI1227 07:43:04.648539 6555 handler.go:208] Removed *v1.Node event handler 7\\\\nI1227 07:43:04.648553 6555 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1227 07:43:04.648568 6555 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1227 07:43:04.649724 6555 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1227 07:43:04.649797 6555 factory.go:656] Stopping watch factory\\\\nI1227 07:43:04.649819 6555 ovnkube.go:599] Stopped ovnkube\\\\nI1227 07:43:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:43:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:29Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.625140 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.625193 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.625205 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.625225 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.625239 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:29Z","lastTransitionTime":"2025-12-27T07:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.628953 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:29Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.644157 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:29Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.656857 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:29Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.674063 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd93f8a550f1648117195c4caac3c1e150c84a1c2dba90ae1ba1a4ecc1bd284b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:25Z\\\",\\\"message\\\":\\\"2025-12-27T07:42:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b83db5b0-0703-4f20-99c4-88b5d3e7ba75\\\\n2025-12-27T07:42:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b83db5b0-0703-4f20-99c4-88b5d3e7ba75 to /host/opt/cni/bin/\\\\n2025-12-27T07:42:40Z [verbose] multus-daemon started\\\\n2025-12-27T07:42:40Z [verbose] Readiness Indicator file check\\\\n2025-12-27T07:43:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:43:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:29Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.690038 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:29Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.702529 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:29Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.714150 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:29Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.726219 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:29Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.727504 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.727551 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.727563 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.727581 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.727593 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:29Z","lastTransitionTime":"2025-12-27T07:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.829755 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.829803 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.829816 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.829835 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.829855 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:29Z","lastTransitionTime":"2025-12-27T07:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.932255 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.932291 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.932299 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.932311 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:29 crc kubenswrapper[4934]: I1227 07:43:29.932321 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:29Z","lastTransitionTime":"2025-12-27T07:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.035948 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.035985 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.035993 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.036008 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.036018 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:30Z","lastTransitionTime":"2025-12-27T07:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.137820 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.137874 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.137892 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.137916 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.137933 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:30Z","lastTransitionTime":"2025-12-27T07:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.240005 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.240047 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.240056 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.240071 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.240097 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:30Z","lastTransitionTime":"2025-12-27T07:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.342331 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.342374 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.342385 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.342400 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.342412 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:30Z","lastTransitionTime":"2025-12-27T07:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.444877 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.444912 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.444924 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.444939 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.444949 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:30Z","lastTransitionTime":"2025-12-27T07:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.466610 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:30 crc kubenswrapper[4934]: E1227 07:43:30.466776 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.547398 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.547436 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.547448 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.547464 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.547476 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:30Z","lastTransitionTime":"2025-12-27T07:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.649928 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.650023 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.650043 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.650067 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.650107 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:30Z","lastTransitionTime":"2025-12-27T07:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.752363 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.752446 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.752471 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.752502 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.752527 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:30Z","lastTransitionTime":"2025-12-27T07:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.855277 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.855335 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.855352 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.855376 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.855394 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:30Z","lastTransitionTime":"2025-12-27T07:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.958290 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.958351 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.958371 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.958396 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:30 crc kubenswrapper[4934]: I1227 07:43:30.958413 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:30Z","lastTransitionTime":"2025-12-27T07:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.061681 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.061751 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.061775 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.061802 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.061824 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:31Z","lastTransitionTime":"2025-12-27T07:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.164604 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.164863 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.164935 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.165015 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.165107 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:31Z","lastTransitionTime":"2025-12-27T07:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.268008 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.268067 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.268079 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.268118 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.268129 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:31Z","lastTransitionTime":"2025-12-27T07:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.369921 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.370274 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.370347 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.370418 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.370484 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:31Z","lastTransitionTime":"2025-12-27T07:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.466484 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.466503 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:31 crc kubenswrapper[4934]: E1227 07:43:31.467059 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.466552 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:31 crc kubenswrapper[4934]: E1227 07:43:31.467167 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:31 crc kubenswrapper[4934]: E1227 07:43:31.466940 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.472861 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.472901 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.472917 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.472939 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.472956 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:31Z","lastTransitionTime":"2025-12-27T07:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.576321 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.576382 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.576400 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.576422 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.576439 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:31Z","lastTransitionTime":"2025-12-27T07:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.678851 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.678914 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.678927 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.678944 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.678957 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:31Z","lastTransitionTime":"2025-12-27T07:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.781434 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.781482 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.781497 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.781517 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.781531 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:31Z","lastTransitionTime":"2025-12-27T07:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.884666 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.884761 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.884789 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.884824 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.884852 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:31Z","lastTransitionTime":"2025-12-27T07:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.988069 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.988121 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.988129 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.988145 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:31 crc kubenswrapper[4934]: I1227 07:43:31.988153 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:31Z","lastTransitionTime":"2025-12-27T07:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.090202 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.090241 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.090249 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.090264 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.090273 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:32Z","lastTransitionTime":"2025-12-27T07:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.192166 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.192249 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.192272 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.192305 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.192330 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:32Z","lastTransitionTime":"2025-12-27T07:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.294709 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.294763 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.294782 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.294803 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.294819 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:32Z","lastTransitionTime":"2025-12-27T07:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.396754 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.396812 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.396825 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.396843 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.396854 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:32Z","lastTransitionTime":"2025-12-27T07:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.463714 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.463775 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.463792 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.463818 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.463835 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:32Z","lastTransitionTime":"2025-12-27T07:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.466588 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:32 crc kubenswrapper[4934]: E1227 07:43:32.466750 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:32 crc kubenswrapper[4934]: E1227 07:43:32.481531 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:32Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.487941 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.487990 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.488006 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.488034 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.488165 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:32Z","lastTransitionTime":"2025-12-27T07:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:32 crc kubenswrapper[4934]: E1227 07:43:32.505143 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:32Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.510011 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.510061 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.510077 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.510121 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.510138 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:32Z","lastTransitionTime":"2025-12-27T07:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:32 crc kubenswrapper[4934]: E1227 07:43:32.527308 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:32Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.532488 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.532552 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.532571 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.532593 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.532608 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:32Z","lastTransitionTime":"2025-12-27T07:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:32 crc kubenswrapper[4934]: E1227 07:43:32.551256 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:32Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.557877 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.557938 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.557955 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.557979 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.557995 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:32Z","lastTransitionTime":"2025-12-27T07:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:32 crc kubenswrapper[4934]: E1227 07:43:32.574764 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:32Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:32 crc kubenswrapper[4934]: E1227 07:43:32.574993 4934 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.576860 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.576907 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.576929 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.576951 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.576968 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:32Z","lastTransitionTime":"2025-12-27T07:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.679651 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.679707 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.679716 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.679729 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.679740 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:32Z","lastTransitionTime":"2025-12-27T07:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.781997 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.782054 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.782072 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.782121 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.782140 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:32Z","lastTransitionTime":"2025-12-27T07:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.884977 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.885031 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.885047 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.885072 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.885121 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:32Z","lastTransitionTime":"2025-12-27T07:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.987832 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.987907 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.987924 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.987948 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:32 crc kubenswrapper[4934]: I1227 07:43:32.987965 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:32Z","lastTransitionTime":"2025-12-27T07:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.090432 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.090494 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.090512 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.090548 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.090567 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:33Z","lastTransitionTime":"2025-12-27T07:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.192784 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.192855 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.192876 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.192903 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.192925 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:33Z","lastTransitionTime":"2025-12-27T07:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.295772 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.295819 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.295830 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.295847 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.295860 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:33Z","lastTransitionTime":"2025-12-27T07:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.399115 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.399182 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.399202 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.399228 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.399248 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:33Z","lastTransitionTime":"2025-12-27T07:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.469820 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:33 crc kubenswrapper[4934]: E1227 07:43:33.469934 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.470118 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:33 crc kubenswrapper[4934]: E1227 07:43:33.470181 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.470321 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:33 crc kubenswrapper[4934]: E1227 07:43:33.470382 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.501477 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.501519 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.501531 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.501547 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.501558 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:33Z","lastTransitionTime":"2025-12-27T07:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.604282 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.604320 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.604331 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.604346 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.604357 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:33Z","lastTransitionTime":"2025-12-27T07:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.707866 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.707933 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.707957 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.708002 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.708026 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:33Z","lastTransitionTime":"2025-12-27T07:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.810836 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.810882 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.810897 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.810918 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.810934 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:33Z","lastTransitionTime":"2025-12-27T07:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.914269 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.914325 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.914343 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.914422 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:33 crc kubenswrapper[4934]: I1227 07:43:33.914444 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:33Z","lastTransitionTime":"2025-12-27T07:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.016524 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.016559 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.016570 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.016584 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.016594 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:34Z","lastTransitionTime":"2025-12-27T07:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.119447 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.119495 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.119509 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.119526 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.119538 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:34Z","lastTransitionTime":"2025-12-27T07:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.223387 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.223448 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.223468 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.223492 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.223510 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:34Z","lastTransitionTime":"2025-12-27T07:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.326405 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.326461 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.326479 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.326502 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.326520 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:34Z","lastTransitionTime":"2025-12-27T07:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.430176 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.430232 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.430250 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.430276 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.430294 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:34Z","lastTransitionTime":"2025-12-27T07:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.467049 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:34 crc kubenswrapper[4934]: E1227 07:43:34.467303 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.468693 4934 scope.go:117] "RemoveContainer" containerID="8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.534796 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.535020 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.535274 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.535469 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.535697 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:34Z","lastTransitionTime":"2025-12-27T07:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.638895 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.638938 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.638949 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.638982 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.638997 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:34Z","lastTransitionTime":"2025-12-27T07:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.742018 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.742072 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.742114 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.742137 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.742154 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:34Z","lastTransitionTime":"2025-12-27T07:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.845167 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.845214 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.845226 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.845242 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.845254 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:34Z","lastTransitionTime":"2025-12-27T07:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.947057 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.947127 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.947138 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.947155 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:34 crc kubenswrapper[4934]: I1227 07:43:34.947166 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:34Z","lastTransitionTime":"2025-12-27T07:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.051264 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.051376 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.051399 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.051465 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.051485 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:35Z","lastTransitionTime":"2025-12-27T07:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.155780 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.155825 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.155837 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.155854 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.155869 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:35Z","lastTransitionTime":"2025-12-27T07:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.259539 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.259627 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.259674 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.259697 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.259714 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:35Z","lastTransitionTime":"2025-12-27T07:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.362220 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.362286 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.362305 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.362333 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.362352 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:35Z","lastTransitionTime":"2025-12-27T07:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.464974 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.465037 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.465054 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.465079 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.465121 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:35Z","lastTransitionTime":"2025-12-27T07:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.466903 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.466923 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:35 crc kubenswrapper[4934]: E1227 07:43:35.467033 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.466912 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:35 crc kubenswrapper[4934]: E1227 07:43:35.467202 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:35 crc kubenswrapper[4934]: E1227 07:43:35.467339 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.568024 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.568107 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.568128 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.568154 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.568172 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:35Z","lastTransitionTime":"2025-12-27T07:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.671114 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.671157 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.671168 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.671194 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.671205 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:35Z","lastTransitionTime":"2025-12-27T07:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.774162 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.774251 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.774275 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.774320 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.774344 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:35Z","lastTransitionTime":"2025-12-27T07:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.877145 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.877239 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.877260 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.877284 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.877302 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:35Z","lastTransitionTime":"2025-12-27T07:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.963489 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovnkube-controller/2.log" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.967001 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerStarted","Data":"0133bfa46253ea9cf0c58a1951b5b012f93f5e8e1ae0ce6bff2341e6d94366c4"} Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.967484 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.979820 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.979873 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.979890 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.979916 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.979934 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:35Z","lastTransitionTime":"2025-12-27T07:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.984901 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:35Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:35 crc kubenswrapper[4934]: I1227 07:43:35.999709 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"304709d3-05ab-4d27-8722-f94053159a29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jjlqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:35Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.014580 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ee2490b-5c06-4f10-a362-b5fcb71f32a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80370fd8607ddb05cccbe7383b94c5a4c1a874299bbbc7f665805ae5edf73ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e461c87813b05a23e571f459ad229874b607be5916ab31f502d476655d5f3d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://243b20c55923c82f8d7a735a9bd0656bfe13008645b5b637c9d0f3890b5890d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:36Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.030280 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:36Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.045675 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:36Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.070417 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0133bfa46253ea9cf0c58a1951b5b012f93f5e8e1ae0ce6bff2341e6d94366c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:04Z\\\",\\\"message\\\":\\\":43:04.648259 6555 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1227 07:43:04.648311 6555 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1227 07:43:04.648332 6555 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1227 07:43:04.648368 6555 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1227 07:43:04.648417 6555 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1227 07:43:04.648433 6555 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1227 07:43:04.648472 6555 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1227 07:43:04.648490 6555 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1227 07:43:04.648506 6555 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1227 07:43:04.648525 6555 handler.go:208] Removed *v1.Node event handler 2\\\\nI1227 07:43:04.648539 6555 handler.go:208] Removed *v1.Node event handler 7\\\\nI1227 07:43:04.648553 6555 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1227 07:43:04.648568 6555 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1227 07:43:04.649724 6555 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1227 07:43:04.649797 6555 factory.go:656] Stopping watch factory\\\\nI1227 07:43:04.649819 6555 ovnkube.go:599] Stopped ovnkube\\\\nI1227 07:43:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:43:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:36Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.084282 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.084337 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.084353 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.084376 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.084393 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:36Z","lastTransitionTime":"2025-12-27T07:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.087838 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd63c24f-fd03-4148-a9a0-a5b86882b52a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe31fde827f2497b12876938f30cc22d79ece0e3e2ab57a034266de29d28266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a5ed5db5f0c41fa30232392526be2a603c1fdc408aef12e8296deea70899805\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g2dnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:36Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.106174 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:36Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.119056 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:36Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.134014 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:36Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.148694 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:36Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.165656 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:36Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.179410 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:36Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.186764 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.186793 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.186803 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.186816 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.186827 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:36Z","lastTransitionTime":"2025-12-27T07:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.195466 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd93f8a550f1648117195c4caac3c1e150c84a1c2dba90ae1ba1a4ecc1bd284b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:25Z\\\",\\\"message\\\":\\\"2025-12-27T07:42:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b83db5b0-0703-4f20-99c4-88b5d3e7ba75\\\\n2025-12-27T07:42:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b83db5b0-0703-4f20-99c4-88b5d3e7ba75 to /host/opt/cni/bin/\\\\n2025-12-27T07:42:40Z [verbose] multus-daemon started\\\\n2025-12-27T07:42:40Z [verbose] Readiness Indicator file check\\\\n2025-12-27T07:43:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:43:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:36Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.219810 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:36Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.232176 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:36Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.243798 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:36Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.289274 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.289361 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.289370 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.289382 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.289390 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:36Z","lastTransitionTime":"2025-12-27T07:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.391535 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.391595 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.391615 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.391639 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.391656 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:36Z","lastTransitionTime":"2025-12-27T07:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.466552 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:36 crc kubenswrapper[4934]: E1227 07:43:36.466692 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.494100 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.494126 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.494137 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.494150 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.494161 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:36Z","lastTransitionTime":"2025-12-27T07:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.596835 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.596912 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.596923 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.596938 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.596949 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:36Z","lastTransitionTime":"2025-12-27T07:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.698904 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.698961 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.698974 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.698993 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.699006 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:36Z","lastTransitionTime":"2025-12-27T07:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.820686 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.820715 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.820723 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.820735 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.820744 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:36Z","lastTransitionTime":"2025-12-27T07:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.922987 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.923041 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.923059 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.923117 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.923136 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:36Z","lastTransitionTime":"2025-12-27T07:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.973302 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovnkube-controller/3.log" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.979319 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovnkube-controller/2.log" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.983997 4934 generic.go:334] "Generic (PLEG): container finished" podID="1c4780a7-56b7-4772-af66-045a03b31a71" containerID="0133bfa46253ea9cf0c58a1951b5b012f93f5e8e1ae0ce6bff2341e6d94366c4" exitCode=1 Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.984052 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerDied","Data":"0133bfa46253ea9cf0c58a1951b5b012f93f5e8e1ae0ce6bff2341e6d94366c4"} Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.984142 4934 scope.go:117] "RemoveContainer" containerID="8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05" Dec 27 07:43:36 crc kubenswrapper[4934]: I1227 07:43:36.984851 4934 scope.go:117] "RemoveContainer" containerID="0133bfa46253ea9cf0c58a1951b5b012f93f5e8e1ae0ce6bff2341e6d94366c4" Dec 27 07:43:36 crc kubenswrapper[4934]: E1227 07:43:36.985036 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.007837 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:37Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.025522 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.025558 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.025569 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.025588 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.025600 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:37Z","lastTransitionTime":"2025-12-27T07:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.026448 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:37Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.055592 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0133bfa46253ea9cf0c58a1951b5b012f93f5e8e1ae0ce6bff2341e6d94366c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8e37444a36bb810406befde03171a2e92921a09ab2c285331058f8f744a73d05\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:04Z\\\",\\\"message\\\":\\\":43:04.648259 6555 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1227 07:43:04.648311 6555 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1227 07:43:04.648332 6555 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1227 07:43:04.648368 6555 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1227 07:43:04.648417 6555 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1227 07:43:04.648433 6555 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1227 07:43:04.648472 6555 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1227 07:43:04.648490 6555 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1227 07:43:04.648506 6555 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1227 07:43:04.648525 6555 handler.go:208] Removed *v1.Node event handler 2\\\\nI1227 07:43:04.648539 6555 handler.go:208] Removed *v1.Node event handler 7\\\\nI1227 07:43:04.648553 6555 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1227 07:43:04.648568 6555 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1227 07:43:04.649724 6555 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1227 07:43:04.649797 6555 factory.go:656] Stopping watch factory\\\\nI1227 07:43:04.649819 6555 ovnkube.go:599] Stopped ovnkube\\\\nI1227 07:43:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:43:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0133bfa46253ea9cf0c58a1951b5b012f93f5e8e1ae0ce6bff2341e6d94366c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:36Z\\\",\\\"message\\\":\\\"ClusterIPs:[10.217.4.174],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1227 07:43:35.975250 6962 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1227 07:43:35.975253 6962 lb_config.go:1031] Cluster endpoints for openshift-dns-operator/metrics for network=default are: map[]\\\\nI1227 07:43:35.975261 6962 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1227 07:43:35.975271 6962 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1227 07:43:35.975319 6962 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1227 07:43:35.975347 6962 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1227 07:43:35.975434 6962 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:43:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:37Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.071850 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd63c24f-fd03-4148-a9a0-a5b86882b52a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe31fde827f2497b12876938f30cc22d79ece0e3e2ab57a034266de29d28266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a5ed5db5f0c41fa30232392526be2a603c1fdc408aef12e8296deea70899805\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g2dnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:37Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.095767 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:37Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.114821 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:37Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.129330 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.129563 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.129730 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.129890 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.130015 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:37Z","lastTransitionTime":"2025-12-27T07:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.130787 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:37Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.145374 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:37Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.161268 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:37Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.175389 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:37Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.190990 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:37Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.213541 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd93f8a550f1648117195c4caac3c1e150c84a1c2dba90ae1ba1a4ecc1bd284b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:25Z\\\",\\\"message\\\":\\\"2025-12-27T07:42:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b83db5b0-0703-4f20-99c4-88b5d3e7ba75\\\\n2025-12-27T07:42:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b83db5b0-0703-4f20-99c4-88b5d3e7ba75 to /host/opt/cni/bin/\\\\n2025-12-27T07:42:40Z [verbose] multus-daemon started\\\\n2025-12-27T07:42:40Z [verbose] Readiness Indicator file check\\\\n2025-12-27T07:43:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:43:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:37Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.230883 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ee2490b-5c06-4f10-a362-b5fcb71f32a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80370fd8607ddb05cccbe7383b94c5a4c1a874299bbbc7f665805ae5edf73ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e461c87813b05a23e571f459ad229874b607be5916ab31f502d476655d5f3d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://243b20c55923c82f8d7a735a9bd0656bfe13008645b5b637c9d0f3890b5890d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:37Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.232824 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.232874 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.232891 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.232916 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.232934 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:37Z","lastTransitionTime":"2025-12-27T07:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.250921 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:37Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.274726 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:37Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.292866 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:37Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.309906 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"304709d3-05ab-4d27-8722-f94053159a29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jjlqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:37Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.335469 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.335513 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.335557 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.335577 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.335592 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:37Z","lastTransitionTime":"2025-12-27T07:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.438403 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.438434 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.438444 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.438461 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.438471 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:37Z","lastTransitionTime":"2025-12-27T07:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.466774 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:37 crc kubenswrapper[4934]: E1227 07:43:37.466882 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.466939 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:37 crc kubenswrapper[4934]: E1227 07:43:37.467145 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.467157 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:37 crc kubenswrapper[4934]: E1227 07:43:37.467281 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.540841 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.540916 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.540940 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.540969 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.540991 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:37Z","lastTransitionTime":"2025-12-27T07:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.644040 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.644122 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.644141 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.644165 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.644210 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:37Z","lastTransitionTime":"2025-12-27T07:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.747475 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.747537 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.747556 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.747580 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.747598 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:37Z","lastTransitionTime":"2025-12-27T07:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.850296 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.850346 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.850354 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.850370 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.850379 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:37Z","lastTransitionTime":"2025-12-27T07:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.953405 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.953454 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.953468 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.953486 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.953499 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:37Z","lastTransitionTime":"2025-12-27T07:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:37 crc kubenswrapper[4934]: I1227 07:43:37.989706 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovnkube-controller/3.log" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:37.993848 4934 scope.go:117] "RemoveContainer" containerID="0133bfa46253ea9cf0c58a1951b5b012f93f5e8e1ae0ce6bff2341e6d94366c4" Dec 27 07:43:38 crc kubenswrapper[4934]: E1227 07:43:37.994016 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.008543 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.023644 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.035974 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.048190 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"304709d3-05ab-4d27-8722-f94053159a29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jjlqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.056311 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.056368 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.056379 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.056396 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.056407 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:38Z","lastTransitionTime":"2025-12-27T07:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.064320 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ee2490b-5c06-4f10-a362-b5fcb71f32a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80370fd8607ddb05cccbe7383b94c5a4c1a874299bbbc7f665805ae5edf73ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e461c87813b05a23e571f459ad229874b607be5916ab31f502d476655d5f3d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://243b20c55923c82f8d7a735a9bd0656bfe13008645b5b637c9d0f3890b5890d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.078017 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.092335 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.110072 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0133bfa46253ea9cf0c58a1951b5b012f93f5e8e1ae0ce6bff2341e6d94366c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0133bfa46253ea9cf0c58a1951b5b012f93f5e8e1ae0ce6bff2341e6d94366c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:36Z\\\",\\\"message\\\":\\\"ClusterIPs:[10.217.4.174],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1227 07:43:35.975250 6962 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1227 07:43:35.975253 6962 lb_config.go:1031] Cluster endpoints for openshift-dns-operator/metrics for network=default are: map[]\\\\nI1227 07:43:35.975261 6962 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1227 07:43:35.975271 6962 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1227 07:43:35.975319 6962 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1227 07:43:35.975347 6962 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1227 07:43:35.975434 6962 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:43:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.122756 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd63c24f-fd03-4148-a9a0-a5b86882b52a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe31fde827f2497b12876938f30cc22d79ece0e3e2ab57a034266de29d28266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a5ed5db5f0c41fa30232392526be2a603c1fdc408aef12e8296deea70899805\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g2dnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.138038 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.150708 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.158326 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.158399 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.158425 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.158449 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.158472 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:38Z","lastTransitionTime":"2025-12-27T07:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.165517 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.176638 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.186773 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.197705 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.209017 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd93f8a550f1648117195c4caac3c1e150c84a1c2dba90ae1ba1a4ecc1bd284b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:25Z\\\",\\\"message\\\":\\\"2025-12-27T07:42:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b83db5b0-0703-4f20-99c4-88b5d3e7ba75\\\\n2025-12-27T07:42:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b83db5b0-0703-4f20-99c4-88b5d3e7ba75 to /host/opt/cni/bin/\\\\n2025-12-27T07:42:40Z [verbose] multus-daemon started\\\\n2025-12-27T07:42:40Z [verbose] Readiness Indicator file check\\\\n2025-12-27T07:43:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:43:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.223467 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:38Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.260869 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.260917 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.260928 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.260945 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.260959 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:38Z","lastTransitionTime":"2025-12-27T07:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.363898 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.363953 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.363970 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.363993 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.364010 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:38Z","lastTransitionTime":"2025-12-27T07:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.466170 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.466232 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.466253 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.466282 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.466303 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:38Z","lastTransitionTime":"2025-12-27T07:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.466449 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:38 crc kubenswrapper[4934]: E1227 07:43:38.466638 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.569129 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.569158 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.569165 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.569177 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.569186 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:38Z","lastTransitionTime":"2025-12-27T07:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.671777 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.671846 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.671863 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.671885 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.671902 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:38Z","lastTransitionTime":"2025-12-27T07:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.774727 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.774779 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.774796 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.774816 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.774830 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:38Z","lastTransitionTime":"2025-12-27T07:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.878259 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.878306 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.878324 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.878348 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.878365 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:38Z","lastTransitionTime":"2025-12-27T07:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.980305 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.980374 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.980408 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.980445 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:38 crc kubenswrapper[4934]: I1227 07:43:38.980465 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:38Z","lastTransitionTime":"2025-12-27T07:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.083527 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.083580 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.083593 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.083622 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.083642 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:39Z","lastTransitionTime":"2025-12-27T07:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.186415 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.186451 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.186462 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.186475 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.186484 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:39Z","lastTransitionTime":"2025-12-27T07:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.290836 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.290941 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.290995 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.291020 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.291067 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:39Z","lastTransitionTime":"2025-12-27T07:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.394324 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.394385 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.394406 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.394432 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.394453 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:39Z","lastTransitionTime":"2025-12-27T07:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.467527 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:39 crc kubenswrapper[4934]: E1227 07:43:39.467665 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.467900 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:39 crc kubenswrapper[4934]: E1227 07:43:39.467988 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.468353 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:39 crc kubenswrapper[4934]: E1227 07:43:39.468462 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.483628 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.497296 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.497400 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.497418 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.497443 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.497460 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:39Z","lastTransitionTime":"2025-12-27T07:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.501633 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.521486 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd93f8a550f1648117195c4caac3c1e150c84a1c2dba90ae1ba1a4ecc1bd284b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:25Z\\\",\\\"message\\\":\\\"2025-12-27T07:42:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b83db5b0-0703-4f20-99c4-88b5d3e7ba75\\\\n2025-12-27T07:42:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b83db5b0-0703-4f20-99c4-88b5d3e7ba75 to /host/opt/cni/bin/\\\\n2025-12-27T07:42:40Z [verbose] multus-daemon started\\\\n2025-12-27T07:42:40Z [verbose] Readiness Indicator file check\\\\n2025-12-27T07:43:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:43:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.541858 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.561937 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.579586 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.592983 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.599368 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.599432 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.599449 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.599468 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.599509 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:39Z","lastTransitionTime":"2025-12-27T07:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.609093 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"304709d3-05ab-4d27-8722-f94053159a29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jjlqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.625087 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ee2490b-5c06-4f10-a362-b5fcb71f32a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80370fd8607ddb05cccbe7383b94c5a4c1a874299bbbc7f665805ae5edf73ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e461c87813b05a23e571f459ad229874b607be5916ab31f502d476655d5f3d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://243b20c55923c82f8d7a735a9bd0656bfe13008645b5b637c9d0f3890b5890d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.635414 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.647494 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.665332 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0133bfa46253ea9cf0c58a1951b5b012f93f5e8e1ae0ce6bff2341e6d94366c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0133bfa46253ea9cf0c58a1951b5b012f93f5e8e1ae0ce6bff2341e6d94366c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:36Z\\\",\\\"message\\\":\\\"ClusterIPs:[10.217.4.174],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1227 07:43:35.975250 6962 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1227 07:43:35.975253 6962 lb_config.go:1031] Cluster endpoints for openshift-dns-operator/metrics for network=default are: map[]\\\\nI1227 07:43:35.975261 6962 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1227 07:43:35.975271 6962 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1227 07:43:35.975319 6962 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1227 07:43:35.975347 6962 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1227 07:43:35.975434 6962 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:43:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.675339 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd63c24f-fd03-4148-a9a0-a5b86882b52a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe31fde827f2497b12876938f30cc22d79ece0e3e2ab57a034266de29d28266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a5ed5db5f0c41fa30232392526be2a603c1fdc408aef12e8296deea70899805\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g2dnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.686074 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.699414 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.702660 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.702694 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.702703 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.702716 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.702726 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:39Z","lastTransitionTime":"2025-12-27T07:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.713148 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.728757 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:39Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.805152 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.805247 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.805267 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.805293 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.805315 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:39Z","lastTransitionTime":"2025-12-27T07:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.908693 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.908725 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.908733 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.908745 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:39 crc kubenswrapper[4934]: I1227 07:43:39.908754 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:39Z","lastTransitionTime":"2025-12-27T07:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.011855 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.011912 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.011934 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.011961 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.011983 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:40Z","lastTransitionTime":"2025-12-27T07:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.115127 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.115210 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.115234 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.115263 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.115285 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:40Z","lastTransitionTime":"2025-12-27T07:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.217923 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.218017 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.218039 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.218061 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.218077 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:40Z","lastTransitionTime":"2025-12-27T07:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.320117 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.320189 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.320212 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.320239 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.320259 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:40Z","lastTransitionTime":"2025-12-27T07:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.423065 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.423220 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.423247 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.423275 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.423294 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:40Z","lastTransitionTime":"2025-12-27T07:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.467269 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:40 crc kubenswrapper[4934]: E1227 07:43:40.467473 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.525684 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.525774 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.525795 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.525822 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.525839 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:40Z","lastTransitionTime":"2025-12-27T07:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.628721 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.628790 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.628812 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.628839 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.628860 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:40Z","lastTransitionTime":"2025-12-27T07:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.732524 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.732599 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.732621 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.732651 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.732672 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:40Z","lastTransitionTime":"2025-12-27T07:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.836304 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.836455 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.836484 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.836515 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.836536 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:40Z","lastTransitionTime":"2025-12-27T07:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.939851 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.939952 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.940001 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.940028 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:40 crc kubenswrapper[4934]: I1227 07:43:40.940070 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:40Z","lastTransitionTime":"2025-12-27T07:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.043516 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.043946 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.043967 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.043993 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.044012 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:41Z","lastTransitionTime":"2025-12-27T07:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.147145 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.147222 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.147251 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.147284 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.147308 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:41Z","lastTransitionTime":"2025-12-27T07:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.205225 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:43:41 crc kubenswrapper[4934]: E1227 07:43:41.205642 4934 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.206020 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:41 crc kubenswrapper[4934]: E1227 07:43:41.206182 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:45.206120974 +0000 UTC m=+146.026561598 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:43:41 crc kubenswrapper[4934]: E1227 07:43:41.206302 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-27 07:44:45.206287898 +0000 UTC m=+146.026728532 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.206406 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:41 crc kubenswrapper[4934]: E1227 07:43:41.206575 4934 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 27 07:43:41 crc kubenswrapper[4934]: E1227 07:43:41.206781 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-27 07:44:45.206766062 +0000 UTC m=+146.027206686 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.251175 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.251298 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.251653 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.251682 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.252180 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:41Z","lastTransitionTime":"2025-12-27T07:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.308584 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:41 crc kubenswrapper[4934]: E1227 07:43:41.308805 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 27 07:43:41 crc kubenswrapper[4934]: E1227 07:43:41.308831 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 27 07:43:41 crc kubenswrapper[4934]: E1227 07:43:41.308850 4934 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.308802 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:41 crc kubenswrapper[4934]: E1227 07:43:41.308919 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-27 07:44:45.308896472 +0000 UTC m=+146.129337106 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:43:41 crc kubenswrapper[4934]: E1227 07:43:41.308943 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 27 07:43:41 crc kubenswrapper[4934]: E1227 07:43:41.309142 4934 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 27 07:43:41 crc kubenswrapper[4934]: E1227 07:43:41.309170 4934 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:43:41 crc kubenswrapper[4934]: E1227 07:43:41.309251 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-27 07:44:45.309229771 +0000 UTC m=+146.129670375 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.355162 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.355388 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.355602 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.355799 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.355971 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:41Z","lastTransitionTime":"2025-12-27T07:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.460029 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.460127 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.460145 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.460176 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.460193 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:41Z","lastTransitionTime":"2025-12-27T07:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.467729 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.467749 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.468218 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:41 crc kubenswrapper[4934]: E1227 07:43:41.468937 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:41 crc kubenswrapper[4934]: E1227 07:43:41.469634 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:41 crc kubenswrapper[4934]: E1227 07:43:41.470004 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.563310 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.563382 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.563424 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.563456 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.563480 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:41Z","lastTransitionTime":"2025-12-27T07:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.666887 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.666941 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.666960 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.666982 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.666999 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:41Z","lastTransitionTime":"2025-12-27T07:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.770826 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.770896 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.770922 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.770950 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.770972 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:41Z","lastTransitionTime":"2025-12-27T07:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.873414 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.873750 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.873915 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.874089 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.874271 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:41Z","lastTransitionTime":"2025-12-27T07:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.977618 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.977684 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.977701 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.977723 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:41 crc kubenswrapper[4934]: I1227 07:43:41.977740 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:41Z","lastTransitionTime":"2025-12-27T07:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.080591 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.080973 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.081213 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.081415 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.081595 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:42Z","lastTransitionTime":"2025-12-27T07:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.185145 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.185903 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.185965 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.185998 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.186022 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:42Z","lastTransitionTime":"2025-12-27T07:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.289328 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.289424 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.289442 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.289494 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.289517 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:42Z","lastTransitionTime":"2025-12-27T07:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.392709 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.392768 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.392786 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.392813 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.392831 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:42Z","lastTransitionTime":"2025-12-27T07:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.467325 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:42 crc kubenswrapper[4934]: E1227 07:43:42.467587 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.495400 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.495453 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.495468 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.495492 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.495507 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:42Z","lastTransitionTime":"2025-12-27T07:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.598945 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.599220 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.599299 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.599382 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.599464 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:42Z","lastTransitionTime":"2025-12-27T07:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.638798 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.638922 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.638993 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.639060 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.639138 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:42Z","lastTransitionTime":"2025-12-27T07:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:42 crc kubenswrapper[4934]: E1227 07:43:42.659144 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.662897 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.662943 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.662975 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.662993 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.663007 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:42Z","lastTransitionTime":"2025-12-27T07:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:42 crc kubenswrapper[4934]: E1227 07:43:42.675714 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.679773 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.679810 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.679826 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.679842 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.679853 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:42Z","lastTransitionTime":"2025-12-27T07:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:42 crc kubenswrapper[4934]: E1227 07:43:42.697850 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.702266 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.702340 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.702360 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.702388 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.702408 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:42Z","lastTransitionTime":"2025-12-27T07:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:42 crc kubenswrapper[4934]: E1227 07:43:42.721392 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.726306 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.726467 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.726544 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.726614 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.726701 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:42Z","lastTransitionTime":"2025-12-27T07:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:42 crc kubenswrapper[4934]: E1227 07:43:42.745509 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:42Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:42 crc kubenswrapper[4934]: E1227 07:43:42.745961 4934 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.749010 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.749137 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.749159 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.749219 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.749251 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:42Z","lastTransitionTime":"2025-12-27T07:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.852594 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.852820 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.852922 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.853052 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.853205 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:42Z","lastTransitionTime":"2025-12-27T07:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.955734 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.955824 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.955835 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.955857 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:42 crc kubenswrapper[4934]: I1227 07:43:42.955872 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:42Z","lastTransitionTime":"2025-12-27T07:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.076422 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.076476 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.076493 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.076516 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.076533 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:43Z","lastTransitionTime":"2025-12-27T07:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.180343 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.180442 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.180493 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.180521 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.180541 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:43Z","lastTransitionTime":"2025-12-27T07:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.284236 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.285283 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.285322 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.285343 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.285354 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:43Z","lastTransitionTime":"2025-12-27T07:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.388613 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.388666 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.388682 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.388708 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.388732 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:43Z","lastTransitionTime":"2025-12-27T07:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.466825 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:43 crc kubenswrapper[4934]: E1227 07:43:43.467068 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.467169 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.467237 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:43 crc kubenswrapper[4934]: E1227 07:43:43.467648 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:43 crc kubenswrapper[4934]: E1227 07:43:43.467704 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.502551 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.504516 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.504688 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.504863 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.504995 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.505238 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:43Z","lastTransitionTime":"2025-12-27T07:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.608193 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.608236 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.608248 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.608263 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.608274 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:43Z","lastTransitionTime":"2025-12-27T07:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.711430 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.711846 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.712018 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.712276 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.712444 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:43Z","lastTransitionTime":"2025-12-27T07:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.815857 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.815904 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.815916 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.815934 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.815946 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:43Z","lastTransitionTime":"2025-12-27T07:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.918723 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.918786 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.918804 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.918829 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:43 crc kubenswrapper[4934]: I1227 07:43:43.918848 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:43Z","lastTransitionTime":"2025-12-27T07:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.021801 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.021856 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.021873 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.021897 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.021915 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:44Z","lastTransitionTime":"2025-12-27T07:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.124796 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.124943 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.124971 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.125001 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.125023 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:44Z","lastTransitionTime":"2025-12-27T07:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.227403 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.227454 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.227473 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.227495 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.227511 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:44Z","lastTransitionTime":"2025-12-27T07:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.330206 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.330275 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.330296 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.330324 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.330345 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:44Z","lastTransitionTime":"2025-12-27T07:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.433594 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.433652 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.433669 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.433691 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.433710 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:44Z","lastTransitionTime":"2025-12-27T07:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.467014 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:44 crc kubenswrapper[4934]: E1227 07:43:44.467291 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.536948 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.537016 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.537039 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.537068 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.537129 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:44Z","lastTransitionTime":"2025-12-27T07:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.640343 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.640399 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.640415 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.640438 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.640455 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:44Z","lastTransitionTime":"2025-12-27T07:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.743291 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.743391 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.743414 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.743442 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.743463 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:44Z","lastTransitionTime":"2025-12-27T07:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.846198 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.846267 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.846284 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.846309 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.846326 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:44Z","lastTransitionTime":"2025-12-27T07:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.949068 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.949204 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.949223 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.949251 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:44 crc kubenswrapper[4934]: I1227 07:43:44.949270 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:44Z","lastTransitionTime":"2025-12-27T07:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.052868 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.052950 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.052997 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.053027 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.053045 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:45Z","lastTransitionTime":"2025-12-27T07:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.156361 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.156424 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.156441 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.156466 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.156484 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:45Z","lastTransitionTime":"2025-12-27T07:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.258797 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.258894 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.258914 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.258940 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.258959 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:45Z","lastTransitionTime":"2025-12-27T07:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.365122 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.365158 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.365169 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.365184 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.365196 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:45Z","lastTransitionTime":"2025-12-27T07:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.466474 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.466576 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.466586 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:45 crc kubenswrapper[4934]: E1227 07:43:45.466670 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:45 crc kubenswrapper[4934]: E1227 07:43:45.466812 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:45 crc kubenswrapper[4934]: E1227 07:43:45.467194 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.468393 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.468435 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.468452 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.468475 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.468493 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:45Z","lastTransitionTime":"2025-12-27T07:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.571419 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.571477 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.571495 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.571523 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.571540 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:45Z","lastTransitionTime":"2025-12-27T07:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.674455 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.674547 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.674572 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.674600 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.674620 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:45Z","lastTransitionTime":"2025-12-27T07:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.777880 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.778363 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.778535 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.778675 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.778795 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:45Z","lastTransitionTime":"2025-12-27T07:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.881639 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.882183 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.882411 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.882555 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.882684 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:45Z","lastTransitionTime":"2025-12-27T07:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.985507 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.985584 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.985611 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.985642 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:45 crc kubenswrapper[4934]: I1227 07:43:45.985666 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:45Z","lastTransitionTime":"2025-12-27T07:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.089489 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.089559 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.089577 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.089602 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.089619 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:46Z","lastTransitionTime":"2025-12-27T07:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.193253 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.193315 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.193330 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.193349 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.193363 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:46Z","lastTransitionTime":"2025-12-27T07:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.296160 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.296217 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.296233 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.296255 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.296271 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:46Z","lastTransitionTime":"2025-12-27T07:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.399032 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.399120 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.399138 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.399161 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.399178 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:46Z","lastTransitionTime":"2025-12-27T07:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.467259 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:46 crc kubenswrapper[4934]: E1227 07:43:46.467845 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.482605 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.501548 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.501619 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.501644 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.501674 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.501696 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:46Z","lastTransitionTime":"2025-12-27T07:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.604171 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.604217 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.604233 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.604255 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.604272 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:46Z","lastTransitionTime":"2025-12-27T07:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.706806 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.706876 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.706899 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.706927 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.706945 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:46Z","lastTransitionTime":"2025-12-27T07:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.809671 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.809735 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.809756 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.809783 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.809802 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:46Z","lastTransitionTime":"2025-12-27T07:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.911512 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.911766 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.911841 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.911932 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:46 crc kubenswrapper[4934]: I1227 07:43:46.912010 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:46Z","lastTransitionTime":"2025-12-27T07:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.015162 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.015244 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.015266 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.015294 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.015314 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:47Z","lastTransitionTime":"2025-12-27T07:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.117751 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.117785 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.117793 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.117807 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.117817 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:47Z","lastTransitionTime":"2025-12-27T07:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.220938 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.221366 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.221576 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.221807 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.222000 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:47Z","lastTransitionTime":"2025-12-27T07:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.324488 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.324551 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.324573 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.324601 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.324622 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:47Z","lastTransitionTime":"2025-12-27T07:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.428273 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.428355 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.428372 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.428437 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.428456 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:47Z","lastTransitionTime":"2025-12-27T07:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.466592 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:47 crc kubenswrapper[4934]: E1227 07:43:47.466786 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.467522 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.467617 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:47 crc kubenswrapper[4934]: E1227 07:43:47.467748 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:47 crc kubenswrapper[4934]: E1227 07:43:47.468040 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.531356 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.531773 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.531916 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.532060 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.532244 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:47Z","lastTransitionTime":"2025-12-27T07:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.635639 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.636009 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.636167 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.636323 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.636461 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:47Z","lastTransitionTime":"2025-12-27T07:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.740352 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.740429 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.740448 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.740472 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.740491 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:47Z","lastTransitionTime":"2025-12-27T07:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.843338 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.843393 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.843410 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.843435 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.843453 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:47Z","lastTransitionTime":"2025-12-27T07:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.946309 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.946531 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.946616 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.946729 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:47 crc kubenswrapper[4934]: I1227 07:43:47.946815 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:47Z","lastTransitionTime":"2025-12-27T07:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.049905 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.049938 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.049947 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.049965 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.049974 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:48Z","lastTransitionTime":"2025-12-27T07:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.152107 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.152355 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.152428 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.152505 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.152563 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:48Z","lastTransitionTime":"2025-12-27T07:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.254297 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.254640 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.254772 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.254862 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.254932 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:48Z","lastTransitionTime":"2025-12-27T07:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.358627 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.358971 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.359171 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.359449 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.359596 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:48Z","lastTransitionTime":"2025-12-27T07:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.462131 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.462201 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.462222 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.462245 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.462262 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:48Z","lastTransitionTime":"2025-12-27T07:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.466403 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:48 crc kubenswrapper[4934]: E1227 07:43:48.466719 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.565019 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.565236 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.565294 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.565350 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.565402 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:48Z","lastTransitionTime":"2025-12-27T07:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.668759 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.668824 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.668842 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.668868 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.668889 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:48Z","lastTransitionTime":"2025-12-27T07:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.772043 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.772181 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.772204 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.772239 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.772264 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:48Z","lastTransitionTime":"2025-12-27T07:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.875532 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.875614 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.875637 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.875670 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.875693 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:48Z","lastTransitionTime":"2025-12-27T07:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.979239 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.979305 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.979329 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.979359 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:48 crc kubenswrapper[4934]: I1227 07:43:48.979379 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:48Z","lastTransitionTime":"2025-12-27T07:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.081543 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.081885 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.082054 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.082230 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.082347 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:49Z","lastTransitionTime":"2025-12-27T07:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.185377 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.185438 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.185450 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.185465 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.185475 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:49Z","lastTransitionTime":"2025-12-27T07:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.288828 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.288892 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.288913 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.288937 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.288956 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:49Z","lastTransitionTime":"2025-12-27T07:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.391315 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.391343 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.391351 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.391364 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.391374 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:49Z","lastTransitionTime":"2025-12-27T07:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.466413 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.466432 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:49 crc kubenswrapper[4934]: E1227 07:43:49.466495 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.466557 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:49 crc kubenswrapper[4934]: E1227 07:43:49.466742 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:49 crc kubenswrapper[4934]: E1227 07:43:49.466843 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.485421 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.493490 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.493534 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.493552 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.493577 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.493598 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:49Z","lastTransitionTime":"2025-12-27T07:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.503240 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.517812 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.535109 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.559007 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd93f8a550f1648117195c4caac3c1e150c84a1c2dba90ae1ba1a4ecc1bd284b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:25Z\\\",\\\"message\\\":\\\"2025-12-27T07:42:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b83db5b0-0703-4f20-99c4-88b5d3e7ba75\\\\n2025-12-27T07:42:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b83db5b0-0703-4f20-99c4-88b5d3e7ba75 to /host/opt/cni/bin/\\\\n2025-12-27T07:42:40Z [verbose] multus-daemon started\\\\n2025-12-27T07:42:40Z [verbose] Readiness Indicator file check\\\\n2025-12-27T07:43:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:43:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.570176 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ee2490b-5c06-4f10-a362-b5fcb71f32a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80370fd8607ddb05cccbe7383b94c5a4c1a874299bbbc7f665805ae5edf73ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e461c87813b05a23e571f459ad229874b607be5916ab31f502d476655d5f3d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://243b20c55923c82f8d7a735a9bd0656bfe13008645b5b637c9d0f3890b5890d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.584733 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.595210 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.595242 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.595292 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.595310 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.595324 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:49Z","lastTransitionTime":"2025-12-27T07:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.599718 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.611749 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.622669 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"304709d3-05ab-4d27-8722-f94053159a29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jjlqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.646302 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb12bd8-afa7-48ff-86f5-064fa312b914\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e4686405be63521dfa03ce0d3940e52a91b7724cc41c18afd134a70f2218985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://999f8cf9a8526d9f8c1f9aa47a2e86b13e4a0fbdfb872b5be86c9a72fa211c2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bbcfef8d68056aba2996d89fcac8d45b1457a76a1fd9320ffd6437a0f7b5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4ef5571db679ca321aed3c4d27699e8ed04433820df8f82089d548b25860d8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34214cc117a89326526885d2667b71811860562c68acc95014c51bbe7996a9af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b202f5c0c838baa712ab5177b345c75f89f955286fbb832c0fa54ad0102646cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b202f5c0c838baa712ab5177b345c75f89f955286fbb832c0fa54ad0102646cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea9ab059166c58229e32334033a120ffde4f8b5212c50c55c97d29eab0f08c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea9ab059166c58229e32334033a120ffde4f8b5212c50c55c97d29eab0f08c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://346ccb1fac116b2783117f09752cc5c013a848dbb02b893ed052989b84af4a81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346ccb1fac116b2783117f09752cc5c013a848dbb02b893ed052989b84af4a81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.659468 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.674312 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.696941 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0133bfa46253ea9cf0c58a1951b5b012f93f5e8e1ae0ce6bff2341e6d94366c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0133bfa46253ea9cf0c58a1951b5b012f93f5e8e1ae0ce6bff2341e6d94366c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:36Z\\\",\\\"message\\\":\\\"ClusterIPs:[10.217.4.174],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1227 07:43:35.975250 6962 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1227 07:43:35.975253 6962 lb_config.go:1031] Cluster endpoints for openshift-dns-operator/metrics for network=default are: map[]\\\\nI1227 07:43:35.975261 6962 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1227 07:43:35.975271 6962 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1227 07:43:35.975319 6962 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1227 07:43:35.975347 6962 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1227 07:43:35.975434 6962 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:43:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.698172 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.698224 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.698242 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.698267 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.698286 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:49Z","lastTransitionTime":"2025-12-27T07:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.711991 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd63c24f-fd03-4148-a9a0-a5b86882b52a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe31fde827f2497b12876938f30cc22d79ece0e3e2ab57a034266de29d28266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a5ed5db5f0c41fa30232392526be2a603c1fdc408aef12e8296deea70899805\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g2dnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.724670 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18c7e094-9196-42e2-b0d0-240a9f198883\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bae82234df2e64ff6d6b36e8a2aa39cf767f34d9804d5cdf2566c1efa23ad5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cdf7e91d7475938502db6b0530157177f1c38196efa5b8821d9dfec51640922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cdf7e91d7475938502db6b0530157177f1c38196efa5b8821d9dfec51640922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.743182 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.761408 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.778479 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:49Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.800991 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.801074 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.801135 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.801159 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.801176 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:49Z","lastTransitionTime":"2025-12-27T07:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.903716 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.903816 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.903834 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.903854 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:49 crc kubenswrapper[4934]: I1227 07:43:49.903871 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:49Z","lastTransitionTime":"2025-12-27T07:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.006015 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.006344 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.006419 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.006491 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.006551 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:50Z","lastTransitionTime":"2025-12-27T07:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.108956 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.109532 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.109728 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.109833 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.109918 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:50Z","lastTransitionTime":"2025-12-27T07:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.213695 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.213793 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.213813 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.213837 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.213855 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:50Z","lastTransitionTime":"2025-12-27T07:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.317335 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.317406 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.317431 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.317469 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.317493 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:50Z","lastTransitionTime":"2025-12-27T07:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.420664 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.420729 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.420747 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.420774 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.420794 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:50Z","lastTransitionTime":"2025-12-27T07:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.467047 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:50 crc kubenswrapper[4934]: E1227 07:43:50.467331 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.468298 4934 scope.go:117] "RemoveContainer" containerID="0133bfa46253ea9cf0c58a1951b5b012f93f5e8e1ae0ce6bff2341e6d94366c4" Dec 27 07:43:50 crc kubenswrapper[4934]: E1227 07:43:50.468550 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.523736 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.524067 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.524298 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.524491 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.524685 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:50Z","lastTransitionTime":"2025-12-27T07:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.628550 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.628611 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.628629 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.628656 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.628675 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:50Z","lastTransitionTime":"2025-12-27T07:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.731270 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.731352 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.731372 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.731396 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.731414 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:50Z","lastTransitionTime":"2025-12-27T07:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.833764 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.833830 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.833843 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.833862 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.833876 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:50Z","lastTransitionTime":"2025-12-27T07:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.936216 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.936299 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.936322 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.936356 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:50 crc kubenswrapper[4934]: I1227 07:43:50.936384 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:50Z","lastTransitionTime":"2025-12-27T07:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.039336 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.039401 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.039418 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.039441 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.039457 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:51Z","lastTransitionTime":"2025-12-27T07:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.142181 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.142242 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.142260 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.142284 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.142306 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:51Z","lastTransitionTime":"2025-12-27T07:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.245417 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.245487 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.245505 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.245534 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.245553 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:51Z","lastTransitionTime":"2025-12-27T07:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.348621 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.348684 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.348700 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.348726 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.348747 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:51Z","lastTransitionTime":"2025-12-27T07:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.451854 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.451935 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.451953 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.451979 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.451997 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:51Z","lastTransitionTime":"2025-12-27T07:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.466956 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.467056 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:51 crc kubenswrapper[4934]: E1227 07:43:51.467144 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:51 crc kubenswrapper[4934]: E1227 07:43:51.467375 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.467222 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:51 crc kubenswrapper[4934]: E1227 07:43:51.467525 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.554743 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.554815 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.554832 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.554857 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.554879 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:51Z","lastTransitionTime":"2025-12-27T07:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.658327 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.658382 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.658403 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.658431 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.658453 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:51Z","lastTransitionTime":"2025-12-27T07:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.761688 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.761778 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.761800 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.761825 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.761841 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:51Z","lastTransitionTime":"2025-12-27T07:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.865403 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.865476 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.865499 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.865528 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.865554 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:51Z","lastTransitionTime":"2025-12-27T07:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.968147 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.968204 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.968227 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.968255 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:51 crc kubenswrapper[4934]: I1227 07:43:51.968275 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:51Z","lastTransitionTime":"2025-12-27T07:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.070942 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.071009 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.071032 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.071063 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.071117 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:52Z","lastTransitionTime":"2025-12-27T07:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.174220 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.174545 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.174731 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.174879 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.175027 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:52Z","lastTransitionTime":"2025-12-27T07:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.277973 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.278028 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.278046 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.278068 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.278140 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:52Z","lastTransitionTime":"2025-12-27T07:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.380587 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.380652 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.380670 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.380696 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.380719 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:52Z","lastTransitionTime":"2025-12-27T07:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.466553 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:52 crc kubenswrapper[4934]: E1227 07:43:52.466774 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.484911 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.484980 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.484999 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.485032 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.485053 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:52Z","lastTransitionTime":"2025-12-27T07:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.589116 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.589247 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.589329 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.589362 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.589385 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:52Z","lastTransitionTime":"2025-12-27T07:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.692458 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.692533 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.692579 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.692625 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.692655 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:52Z","lastTransitionTime":"2025-12-27T07:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.795512 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.795561 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.795575 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.795599 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.795614 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:52Z","lastTransitionTime":"2025-12-27T07:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.898962 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.899023 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.899039 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.899065 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.899116 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:52Z","lastTransitionTime":"2025-12-27T07:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.957586 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.957733 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.957757 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.957786 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.957807 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:52Z","lastTransitionTime":"2025-12-27T07:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:52 crc kubenswrapper[4934]: E1227 07:43:52.979452 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:52Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.985503 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.985554 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.985570 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.985593 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:52 crc kubenswrapper[4934]: I1227 07:43:52.985638 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:52Z","lastTransitionTime":"2025-12-27T07:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:53 crc kubenswrapper[4934]: E1227 07:43:53.006704 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.011909 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.011962 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.011980 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.012004 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.012022 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:53Z","lastTransitionTime":"2025-12-27T07:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:53 crc kubenswrapper[4934]: E1227 07:43:53.032689 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.037676 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.037746 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.037759 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.037777 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.037791 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:53Z","lastTransitionTime":"2025-12-27T07:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:53 crc kubenswrapper[4934]: E1227 07:43:53.057368 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.062624 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.062681 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.062698 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.062727 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.062746 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:53Z","lastTransitionTime":"2025-12-27T07:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:53 crc kubenswrapper[4934]: E1227 07:43:53.084516 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:53Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:53 crc kubenswrapper[4934]: E1227 07:43:53.084628 4934 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.086330 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.086419 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.086429 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.086456 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.086468 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:53Z","lastTransitionTime":"2025-12-27T07:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.190004 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.190079 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.190159 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.190188 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.190242 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:53Z","lastTransitionTime":"2025-12-27T07:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.293516 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.293605 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.293626 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.293653 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.293670 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:53Z","lastTransitionTime":"2025-12-27T07:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.396339 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.396406 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.396427 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.396456 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.396478 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:53Z","lastTransitionTime":"2025-12-27T07:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.466783 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.466880 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:53 crc kubenswrapper[4934]: E1227 07:43:53.466985 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.467108 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:53 crc kubenswrapper[4934]: E1227 07:43:53.467241 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:53 crc kubenswrapper[4934]: E1227 07:43:53.467905 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.498983 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.499015 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.499024 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.499037 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.499045 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:53Z","lastTransitionTime":"2025-12-27T07:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.601590 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.601661 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.601683 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.601709 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.601727 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:53Z","lastTransitionTime":"2025-12-27T07:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.704197 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.704257 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.704271 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.704288 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.704301 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:53Z","lastTransitionTime":"2025-12-27T07:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.807540 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.807590 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.807606 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.807630 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.807647 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:53Z","lastTransitionTime":"2025-12-27T07:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.910054 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.910167 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.910185 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.910211 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:53 crc kubenswrapper[4934]: I1227 07:43:53.910227 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:53Z","lastTransitionTime":"2025-12-27T07:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.013149 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.013553 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.013715 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.013848 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.013991 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:54Z","lastTransitionTime":"2025-12-27T07:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.117605 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.117664 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.117686 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.117714 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.117733 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:54Z","lastTransitionTime":"2025-12-27T07:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.221026 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.221134 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.221202 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.221232 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.221251 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:54Z","lastTransitionTime":"2025-12-27T07:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.324732 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.324772 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.324784 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.324799 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.324811 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:54Z","lastTransitionTime":"2025-12-27T07:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.427881 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.427926 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.427940 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.427958 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.427970 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:54Z","lastTransitionTime":"2025-12-27T07:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.466402 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:54 crc kubenswrapper[4934]: E1227 07:43:54.466609 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.530737 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.530805 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.530823 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.530848 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.530868 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:54Z","lastTransitionTime":"2025-12-27T07:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.634439 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.634496 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.634513 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.634536 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.634554 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:54Z","lastTransitionTime":"2025-12-27T07:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.738226 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.738586 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.738756 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.738898 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.739016 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:54Z","lastTransitionTime":"2025-12-27T07:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.842529 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.842625 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.842655 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.842689 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.842714 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:54Z","lastTransitionTime":"2025-12-27T07:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.946208 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.946274 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.946292 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.946317 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:54 crc kubenswrapper[4934]: I1227 07:43:54.946334 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:54Z","lastTransitionTime":"2025-12-27T07:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.049595 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.049668 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.049689 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.049715 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.049733 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:55Z","lastTransitionTime":"2025-12-27T07:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.154383 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.154445 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.154462 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.154481 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.154497 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:55Z","lastTransitionTime":"2025-12-27T07:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.257056 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.257127 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.257138 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.257159 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.257170 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:55Z","lastTransitionTime":"2025-12-27T07:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.359186 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.359234 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.359251 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.359273 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.359288 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:55Z","lastTransitionTime":"2025-12-27T07:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.461784 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.461838 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.461852 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.461871 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.461885 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:55Z","lastTransitionTime":"2025-12-27T07:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.467356 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.467377 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.467378 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:55 crc kubenswrapper[4934]: E1227 07:43:55.467657 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:55 crc kubenswrapper[4934]: E1227 07:43:55.467744 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:55 crc kubenswrapper[4934]: E1227 07:43:55.467812 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.564221 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.565167 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.565190 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.565210 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.565222 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:55Z","lastTransitionTime":"2025-12-27T07:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.667801 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.667859 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.667877 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.667900 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.667917 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:55Z","lastTransitionTime":"2025-12-27T07:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.771166 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.771227 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.771244 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.771266 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.771282 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:55Z","lastTransitionTime":"2025-12-27T07:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.874169 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.874215 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.874230 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.874252 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.874267 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:55Z","lastTransitionTime":"2025-12-27T07:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.977047 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.977141 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.977159 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.977183 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:55 crc kubenswrapper[4934]: I1227 07:43:55.977200 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:55Z","lastTransitionTime":"2025-12-27T07:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.080300 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.080372 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.080393 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.080421 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.080441 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:56Z","lastTransitionTime":"2025-12-27T07:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.182571 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.182625 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.182641 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.182664 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.182680 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:56Z","lastTransitionTime":"2025-12-27T07:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.285169 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.285237 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.285260 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.285291 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.285313 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:56Z","lastTransitionTime":"2025-12-27T07:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.388586 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.388661 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.388678 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.388705 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.388725 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:56Z","lastTransitionTime":"2025-12-27T07:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.466653 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:56 crc kubenswrapper[4934]: E1227 07:43:56.466844 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.491148 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.491248 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.491271 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.491344 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.491368 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:56Z","lastTransitionTime":"2025-12-27T07:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.594775 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.594839 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.594855 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.594878 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.594896 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:56Z","lastTransitionTime":"2025-12-27T07:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.687432 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs\") pod \"network-metrics-daemon-jjlqg\" (UID: \"304709d3-05ab-4d27-8722-f94053159a29\") " pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:56 crc kubenswrapper[4934]: E1227 07:43:56.687696 4934 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 27 07:43:56 crc kubenswrapper[4934]: E1227 07:43:56.687799 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs podName:304709d3-05ab-4d27-8722-f94053159a29 nodeName:}" failed. No retries permitted until 2025-12-27 07:45:00.687773286 +0000 UTC m=+161.508213920 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs") pod "network-metrics-daemon-jjlqg" (UID: "304709d3-05ab-4d27-8722-f94053159a29") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.697018 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.697127 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.697154 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.697184 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.697205 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:56Z","lastTransitionTime":"2025-12-27T07:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.800657 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.800731 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.800754 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.800783 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.800809 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:56Z","lastTransitionTime":"2025-12-27T07:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.903619 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.903678 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.903700 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.903727 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:56 crc kubenswrapper[4934]: I1227 07:43:56.903751 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:56Z","lastTransitionTime":"2025-12-27T07:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.007036 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.007133 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.007158 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.007186 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.007205 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:57Z","lastTransitionTime":"2025-12-27T07:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.110394 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.110442 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.110460 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.110484 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.110501 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:57Z","lastTransitionTime":"2025-12-27T07:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.212811 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.212870 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.212896 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.212925 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.212947 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:57Z","lastTransitionTime":"2025-12-27T07:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.315895 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.315999 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.316018 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.316043 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.316060 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:57Z","lastTransitionTime":"2025-12-27T07:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.418654 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.418718 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.418736 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.418758 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.418774 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:57Z","lastTransitionTime":"2025-12-27T07:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.466607 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.466727 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:57 crc kubenswrapper[4934]: E1227 07:43:57.466799 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.466903 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:57 crc kubenswrapper[4934]: E1227 07:43:57.467078 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:57 crc kubenswrapper[4934]: E1227 07:43:57.467194 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.521218 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.521275 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.521294 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.521319 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.521336 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:57Z","lastTransitionTime":"2025-12-27T07:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.624469 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.624523 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.624541 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.624563 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.624580 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:57Z","lastTransitionTime":"2025-12-27T07:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.727629 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.727678 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.727721 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.727748 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.727762 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:57Z","lastTransitionTime":"2025-12-27T07:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.831169 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.831220 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.831232 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.831248 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.831259 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:57Z","lastTransitionTime":"2025-12-27T07:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.934422 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.934482 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.934504 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.934531 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:57 crc kubenswrapper[4934]: I1227 07:43:57.934551 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:57Z","lastTransitionTime":"2025-12-27T07:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.037148 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.037214 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.037238 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.037265 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.037286 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:58Z","lastTransitionTime":"2025-12-27T07:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.140154 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.140217 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.140238 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.140269 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.140293 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:58Z","lastTransitionTime":"2025-12-27T07:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.243258 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.243342 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.243359 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.243391 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.243414 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:58Z","lastTransitionTime":"2025-12-27T07:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.346216 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.346284 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.346305 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.346339 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.346361 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:58Z","lastTransitionTime":"2025-12-27T07:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.449333 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.449386 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.449405 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.449425 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.449478 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:58Z","lastTransitionTime":"2025-12-27T07:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.467318 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:43:58 crc kubenswrapper[4934]: E1227 07:43:58.467517 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.552532 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.552590 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.552612 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.552641 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.552665 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:58Z","lastTransitionTime":"2025-12-27T07:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.660540 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.660629 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.660652 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.660679 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.660699 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:58Z","lastTransitionTime":"2025-12-27T07:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.763444 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.763506 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.763523 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.763547 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.763566 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:58Z","lastTransitionTime":"2025-12-27T07:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.866999 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.867061 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.867072 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.867112 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.867122 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:58Z","lastTransitionTime":"2025-12-27T07:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.970126 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.970189 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.970205 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.970230 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:58 crc kubenswrapper[4934]: I1227 07:43:58.970247 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:58Z","lastTransitionTime":"2025-12-27T07:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.072839 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.072898 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.072913 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.072936 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.072955 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:59Z","lastTransitionTime":"2025-12-27T07:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.176581 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.176634 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.176651 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.176673 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.176690 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:59Z","lastTransitionTime":"2025-12-27T07:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.280070 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.280162 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.280180 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.280204 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.280221 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:59Z","lastTransitionTime":"2025-12-27T07:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.384009 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.384103 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.384123 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.384147 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.384164 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:59Z","lastTransitionTime":"2025-12-27T07:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.466512 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.466584 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:43:59 crc kubenswrapper[4934]: E1227 07:43:59.466696 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.466713 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:43:59 crc kubenswrapper[4934]: E1227 07:43:59.466865 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:43:59 crc kubenswrapper[4934]: E1227 07:43:59.466968 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.485405 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c89caceb-57dc-49b0-8e96-69e1ac851beb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://274e17f4ebf530dcdc922fa281ee2bbe94ffae0bc6b25534266ab875721a3781\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://14d24f1d2b53277f5846084c480e28a5e2f31118d56dfcac763b985509e4f93f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e1dc2e113df40d176f488b1b604bca0a1e860ec91142e641da32df38a0ecade3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.488389 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.488470 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.488500 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.488519 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.488531 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:59Z","lastTransitionTime":"2025-12-27T07:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.502771 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a78cd53473610702b08c515e108d77cee83573b3a841d88e813ff3fcc01296c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.512510 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fwcxz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"72743c37-fd5a-4aa8-959b-6a1fd048c650\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9591d5c1bdaa9bdde71cb21759312a880a7382df5b906ee925a9b004f9fdaee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rxfqx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:38Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fwcxz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.526506 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcabb41bc84f64ea0dfe6d6fdb361dca8bc891acca4e68a16464d180f2f0d0ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c572j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-w9j4r\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.539927 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-jzv2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1bdd4b78-f615-4771-9448-8faa03dbc4f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd93f8a550f1648117195c4caac3c1e150c84a1c2dba90ae1ba1a4ecc1bd284b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:25Z\\\",\\\"message\\\":\\\"2025-12-27T07:42:40+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b83db5b0-0703-4f20-99c4-88b5d3e7ba75\\\\n2025-12-27T07:42:40+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b83db5b0-0703-4f20-99c4-88b5d3e7ba75 to /host/opt/cni/bin/\\\\n2025-12-27T07:42:40Z [verbose] multus-daemon started\\\\n2025-12-27T07:42:40Z [verbose] Readiness Indicator file check\\\\n2025-12-27T07:43:25Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:43:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s295j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-jzv2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.551134 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ee2490b-5c06-4f10-a362-b5fcb71f32a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80370fd8607ddb05cccbe7383b94c5a4c1a874299bbbc7f665805ae5edf73ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e461c87813b05a23e571f459ad229874b607be5916ab31f502d476655d5f3d2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://243b20c55923c82f8d7a735a9bd0656bfe13008645b5b637c9d0f3890b5890d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f8b8fff7845567d60d38ec293274997e795e14a276985525863984881fa51912\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.564464 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.580555 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.589906 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.589932 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.589940 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.589953 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.589962 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:59Z","lastTransitionTime":"2025-12-27T07:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.598469 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dfj4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f11f562a-9201-4bd4-9f28-a9b2d5cd44bd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://722e57e68a6ab358bbebb8c3ff371d484d573b010b7a9a6c560839d840d8d79b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k6bzl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dfj4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.611548 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"304709d3-05ab-4d27-8722-f94053159a29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:52Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n4hwf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:52Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-jjlqg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.633101 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bcb12bd8-afa7-48ff-86f5-064fa312b914\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6e4686405be63521dfa03ce0d3940e52a91b7724cc41c18afd134a70f2218985\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://999f8cf9a8526d9f8c1f9aa47a2e86b13e4a0fbdfb872b5be86c9a72fa211c2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2bbcfef8d68056aba2996d89fcac8d45b1457a76a1fd9320ffd6437a0f7b5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4ef5571db679ca321aed3c4d27699e8ed04433820df8f82089d548b25860d8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://34214cc117a89326526885d2667b71811860562c68acc95014c51bbe7996a9af\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b202f5c0c838baa712ab5177b345c75f89f955286fbb832c0fa54ad0102646cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b202f5c0c838baa712ab5177b345c75f89f955286fbb832c0fa54ad0102646cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ea9ab059166c58229e32334033a120ffde4f8b5212c50c55c97d29eab0f08c0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9ea9ab059166c58229e32334033a120ffde4f8b5212c50c55c97d29eab0f08c0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://346ccb1fac116b2783117f09752cc5c013a848dbb02b893ed052989b84af4a81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://346ccb1fac116b2783117f09752cc5c013a848dbb02b893ed052989b84af4a81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.648067 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a41b077f9bda9209323803ed35451b2efc7be55d732a8b796d3f1fcef47ddfd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cbf81b9fbbba45a23a25f904e22e7b5e448f98043db711d2710d25a1f8e3715e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.661487 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e0de454e1d795cb1b05c7a6f263cda34bf19d72de89b0a70e02def0429accdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.688413 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c4780a7-56b7-4772-af66-045a03b31a71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:39Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0133bfa46253ea9cf0c58a1951b5b012f93f5e8e1ae0ce6bff2341e6d94366c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0133bfa46253ea9cf0c58a1951b5b012f93f5e8e1ae0ce6bff2341e6d94366c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-27T07:43:36Z\\\",\\\"message\\\":\\\"ClusterIPs:[10.217.4.174],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1227 07:43:35.975250 6962 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1227 07:43:35.975253 6962 lb_config.go:1031] Cluster endpoints for openshift-dns-operator/metrics for network=default are: map[]\\\\nI1227 07:43:35.975261 6962 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1227 07:43:35.975271 6962 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1227 07:43:35.975319 6962 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nI1227 07:43:35.975347 6962 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1227 07:43:35.975434 6962 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-27T07:43:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2kp9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8vzdl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.693180 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.693217 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.693229 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.693245 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.693256 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:59Z","lastTransitionTime":"2025-12-27T07:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.708622 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cd63c24f-fd03-4148-a9a0-a5b86882b52a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbe31fde827f2497b12876938f30cc22d79ece0e3e2ab57a034266de29d28266\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a5ed5db5f0c41fa30232392526be2a603c1fdc408aef12e8296deea70899805\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v6jxg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:51Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g2dnl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.725969 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18c7e094-9196-42e2-b0d0-240a9f198883\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1bae82234df2e64ff6d6b36e8a2aa39cf767f34d9804d5cdf2566c1efa23ad5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cdf7e91d7475938502db6b0530157177f1c38196efa5b8821d9dfec51640922\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1cdf7e91d7475938502db6b0530157177f1c38196efa5b8821d9dfec51640922\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.748866 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"493e2dd1-d571-4aae-8ce2-132231ee766d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:21Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:19Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.768698 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:37Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.793033 4934 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e59598db-f0ef-42fb-b151-b436ac250b08\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-27T07:42:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f428797c39ee62de08703212693acd757637411265d6e8538c8705aff78cfe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-27T07:42:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5611c491f6839242a80fdb5a28d9482c341b3cbac287db18ccaf669daf9c2692\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://170a5e0d3d788f8e65d7e16e37d09a2a506b426aedc9440e668061f441f1b1d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5aa3439da962ff4b6488047189df36060a81e625b886f5670a1dad3050df2e66\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://667f30ee1039064e281fd03e65d7e4ab758ebf1086f402a1697f37961689c874\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d349657bc57eae5cd0a52a61f5227b3d3bea9c67d2af3373975c61654577d24d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b38904fe2aedbebff8ce468db49356b55e2e11965917727ed3165421809b11db\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-27T07:42:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-27T07:42:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r4ktc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-27T07:42:39Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-hqfw4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:43:59Z is after 2025-08-24T17:21:41Z" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.795639 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.795685 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.795696 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.795717 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.795728 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:59Z","lastTransitionTime":"2025-12-27T07:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.898234 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.898321 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.898343 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.898372 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:43:59 crc kubenswrapper[4934]: I1227 07:43:59.898391 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:43:59Z","lastTransitionTime":"2025-12-27T07:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.001537 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.001601 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.001648 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.001673 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.001691 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:00Z","lastTransitionTime":"2025-12-27T07:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.104390 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.104448 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.104467 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.104491 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.104509 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:00Z","lastTransitionTime":"2025-12-27T07:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.208252 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.208327 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.208345 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.208373 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.208391 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:00Z","lastTransitionTime":"2025-12-27T07:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.311690 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.311749 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.311766 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.311790 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.311808 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:00Z","lastTransitionTime":"2025-12-27T07:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.414485 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.414599 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.414623 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.414652 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.414673 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:00Z","lastTransitionTime":"2025-12-27T07:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.467189 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:44:00 crc kubenswrapper[4934]: E1227 07:44:00.467411 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.517548 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.517625 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.517648 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.517680 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.517704 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:00Z","lastTransitionTime":"2025-12-27T07:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.621117 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.621183 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.621200 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.621228 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.621251 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:00Z","lastTransitionTime":"2025-12-27T07:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.723823 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.723898 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.723920 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.723956 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.723984 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:00Z","lastTransitionTime":"2025-12-27T07:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.827193 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.827249 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.827269 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.827293 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.827311 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:00Z","lastTransitionTime":"2025-12-27T07:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.929204 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.929256 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.929276 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.929298 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:00 crc kubenswrapper[4934]: I1227 07:44:00.929316 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:00Z","lastTransitionTime":"2025-12-27T07:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.033167 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.033363 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.033397 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.033476 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.033502 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:01Z","lastTransitionTime":"2025-12-27T07:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.135955 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.136022 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.136119 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.136154 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.136175 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:01Z","lastTransitionTime":"2025-12-27T07:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.238789 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.239040 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.239246 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.239410 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.239552 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:01Z","lastTransitionTime":"2025-12-27T07:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.342206 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.342504 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.342663 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.342811 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.342979 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:01Z","lastTransitionTime":"2025-12-27T07:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.445940 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.446319 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.446496 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.446631 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.446756 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:01Z","lastTransitionTime":"2025-12-27T07:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.467028 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:01 crc kubenswrapper[4934]: E1227 07:44:01.467223 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.467279 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.467483 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:01 crc kubenswrapper[4934]: E1227 07:44:01.467604 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:44:01 crc kubenswrapper[4934]: E1227 07:44:01.467475 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.475203 4934 scope.go:117] "RemoveContainer" containerID="0133bfa46253ea9cf0c58a1951b5b012f93f5e8e1ae0ce6bff2341e6d94366c4" Dec 27 07:44:01 crc kubenswrapper[4934]: E1227 07:44:01.476758 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.549319 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.549637 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.549853 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.550059 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.550326 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:01Z","lastTransitionTime":"2025-12-27T07:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.653492 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.653569 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.653591 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.653625 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.653652 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:01Z","lastTransitionTime":"2025-12-27T07:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.756213 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.756266 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.756282 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.756305 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.756323 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:01Z","lastTransitionTime":"2025-12-27T07:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.858829 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.858886 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.858905 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.858925 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.858941 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:01Z","lastTransitionTime":"2025-12-27T07:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.961881 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.961956 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.961972 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.961994 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:01 crc kubenswrapper[4934]: I1227 07:44:01.962012 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:01Z","lastTransitionTime":"2025-12-27T07:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.064469 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.064569 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.064589 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.064613 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.064631 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:02Z","lastTransitionTime":"2025-12-27T07:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.167195 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.167541 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.167695 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.167831 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.167962 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:02Z","lastTransitionTime":"2025-12-27T07:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.271563 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.272073 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.272250 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.272412 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.272542 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:02Z","lastTransitionTime":"2025-12-27T07:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.375848 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.375901 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.375917 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.375940 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.375957 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:02Z","lastTransitionTime":"2025-12-27T07:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.467339 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:44:02 crc kubenswrapper[4934]: E1227 07:44:02.467755 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.478521 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.478566 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.478577 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.478651 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.478663 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:02Z","lastTransitionTime":"2025-12-27T07:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.581611 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.581678 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.581696 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.581723 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.581746 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:02Z","lastTransitionTime":"2025-12-27T07:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.684885 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.684961 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.684984 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.685014 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.685033 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:02Z","lastTransitionTime":"2025-12-27T07:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.788203 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.788263 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.788279 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.788301 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.788320 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:02Z","lastTransitionTime":"2025-12-27T07:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.891942 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.892030 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.892056 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.892123 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.892147 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:02Z","lastTransitionTime":"2025-12-27T07:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.994909 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.994995 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.995017 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.995041 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:02 crc kubenswrapper[4934]: I1227 07:44:02.995058 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:02Z","lastTransitionTime":"2025-12-27T07:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.097930 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.097993 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.098016 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.098059 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.098117 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:03Z","lastTransitionTime":"2025-12-27T07:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.201062 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.201147 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.201164 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.201185 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.201202 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:03Z","lastTransitionTime":"2025-12-27T07:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.304258 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.304334 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.304346 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.304392 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.304407 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:03Z","lastTransitionTime":"2025-12-27T07:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.407262 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.407318 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.407335 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.407359 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.407376 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:03Z","lastTransitionTime":"2025-12-27T07:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.418900 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.418966 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.418984 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.419009 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.419027 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:03Z","lastTransitionTime":"2025-12-27T07:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:03 crc kubenswrapper[4934]: E1227 07:44:03.439641 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:44:03Z is after 2025-08-24T17:21:41Z" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.444647 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.444705 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.444727 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.444751 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.444770 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:03Z","lastTransitionTime":"2025-12-27T07:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:03 crc kubenswrapper[4934]: E1227 07:44:03.462992 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:44:03Z is after 2025-08-24T17:21:41Z" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.467041 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.467185 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.467051 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.467464 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.467537 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.467556 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:03 crc kubenswrapper[4934]: E1227 07:44:03.467457 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.467579 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.467598 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:03Z","lastTransitionTime":"2025-12-27T07:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:03 crc kubenswrapper[4934]: E1227 07:44:03.467609 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:44:03 crc kubenswrapper[4934]: E1227 07:44:03.467749 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:44:03 crc kubenswrapper[4934]: E1227 07:44:03.487625 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:44:03Z is after 2025-08-24T17:21:41Z" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.492808 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.492865 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.492884 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.492908 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.492925 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:03Z","lastTransitionTime":"2025-12-27T07:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:03 crc kubenswrapper[4934]: E1227 07:44:03.513826 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:44:03Z is after 2025-08-24T17:21:41Z" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.519032 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.519263 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.519402 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.519525 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.519640 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:03Z","lastTransitionTime":"2025-12-27T07:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:03 crc kubenswrapper[4934]: E1227 07:44:03.539530 4934 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-27T07:44:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"27320adc-74b9-4bb8-b209-d917a73b2290\\\",\\\"systemUUID\\\":\\\"e8025479-33d0-4936-ac24-c37235bb6b99\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-27T07:44:03Z is after 2025-08-24T17:21:41Z" Dec 27 07:44:03 crc kubenswrapper[4934]: E1227 07:44:03.540060 4934 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.542326 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.542384 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.542402 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.542427 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.542445 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:03Z","lastTransitionTime":"2025-12-27T07:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.645545 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.645608 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.645626 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.645651 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.645672 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:03Z","lastTransitionTime":"2025-12-27T07:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.748909 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.748982 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.749005 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.749033 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.749056 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:03Z","lastTransitionTime":"2025-12-27T07:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.852247 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.852348 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.852371 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.852400 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.852422 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:03Z","lastTransitionTime":"2025-12-27T07:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.955670 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.955735 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.955751 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.955779 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:03 crc kubenswrapper[4934]: I1227 07:44:03.955798 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:03Z","lastTransitionTime":"2025-12-27T07:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.058017 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.058114 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.058132 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.058157 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.058176 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:04Z","lastTransitionTime":"2025-12-27T07:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.160874 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.160940 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.160957 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.160982 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.161001 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:04Z","lastTransitionTime":"2025-12-27T07:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.265341 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.265414 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.265443 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.265477 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.265499 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:04Z","lastTransitionTime":"2025-12-27T07:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.368513 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.368816 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.368942 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.369079 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.369278 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:04Z","lastTransitionTime":"2025-12-27T07:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.466724 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:44:04 crc kubenswrapper[4934]: E1227 07:44:04.467256 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.472255 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.472310 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.472329 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.472352 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.472369 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:04Z","lastTransitionTime":"2025-12-27T07:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.574987 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.575047 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.575065 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.575107 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.575125 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:04Z","lastTransitionTime":"2025-12-27T07:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.678302 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.678360 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.678376 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.678401 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.678418 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:04Z","lastTransitionTime":"2025-12-27T07:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.781253 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.781302 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.781322 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.781345 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.781362 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:04Z","lastTransitionTime":"2025-12-27T07:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.883888 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.883932 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.883947 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.883969 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.883986 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:04Z","lastTransitionTime":"2025-12-27T07:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.987227 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.987304 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.987327 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.987362 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:04 crc kubenswrapper[4934]: I1227 07:44:04.987384 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:04Z","lastTransitionTime":"2025-12-27T07:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.090254 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.090582 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.090875 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.091141 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.091369 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:05Z","lastTransitionTime":"2025-12-27T07:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.194496 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.194835 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.195069 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.195340 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.195520 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:05Z","lastTransitionTime":"2025-12-27T07:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.298534 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.298986 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.299276 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.299477 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.299666 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:05Z","lastTransitionTime":"2025-12-27T07:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.403057 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.403209 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.403236 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.403267 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.403291 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:05Z","lastTransitionTime":"2025-12-27T07:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.467324 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.467467 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.467344 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:44:05 crc kubenswrapper[4934]: E1227 07:44:05.467596 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:44:05 crc kubenswrapper[4934]: E1227 07:44:05.467679 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:44:05 crc kubenswrapper[4934]: E1227 07:44:05.467801 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.506885 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.507223 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.507473 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.507649 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.507788 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:05Z","lastTransitionTime":"2025-12-27T07:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.646121 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.646183 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.646196 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.646217 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.646231 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:05Z","lastTransitionTime":"2025-12-27T07:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.749482 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.749541 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.749559 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.749582 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.749600 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:05Z","lastTransitionTime":"2025-12-27T07:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.852844 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.852896 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.852914 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.852937 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.852957 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:05Z","lastTransitionTime":"2025-12-27T07:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.955885 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.955967 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.955991 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.956023 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:05 crc kubenswrapper[4934]: I1227 07:44:05.956050 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:05Z","lastTransitionTime":"2025-12-27T07:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.059518 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.059612 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.059632 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.059665 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.059688 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:06Z","lastTransitionTime":"2025-12-27T07:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.164016 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.164138 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.164227 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.164260 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.164282 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:06Z","lastTransitionTime":"2025-12-27T07:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.268199 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.268250 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.268267 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.268291 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.268333 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:06Z","lastTransitionTime":"2025-12-27T07:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.371278 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.371340 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.371362 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.371391 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.371411 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:06Z","lastTransitionTime":"2025-12-27T07:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.466906 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:44:06 crc kubenswrapper[4934]: E1227 07:44:06.467134 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.474386 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.474629 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.474841 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.475056 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.475325 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:06Z","lastTransitionTime":"2025-12-27T07:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.578675 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.578748 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.578772 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.578800 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.578823 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:06Z","lastTransitionTime":"2025-12-27T07:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.681629 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.681700 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.681716 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.681742 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.681759 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:06Z","lastTransitionTime":"2025-12-27T07:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.784727 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.784763 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.784771 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.784785 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.784793 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:06Z","lastTransitionTime":"2025-12-27T07:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.888677 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.888726 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.888743 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.888766 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.888784 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:06Z","lastTransitionTime":"2025-12-27T07:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.992161 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.992221 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.992240 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.992263 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:06 crc kubenswrapper[4934]: I1227 07:44:06.992281 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:06Z","lastTransitionTime":"2025-12-27T07:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.095257 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.095308 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.095324 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.095349 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.095366 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:07Z","lastTransitionTime":"2025-12-27T07:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.198511 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.198573 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.198592 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.198617 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.198635 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:07Z","lastTransitionTime":"2025-12-27T07:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.302127 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.302178 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.302195 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.302219 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.302237 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:07Z","lastTransitionTime":"2025-12-27T07:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.405440 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.405515 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.405534 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.405561 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.405579 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:07Z","lastTransitionTime":"2025-12-27T07:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.466642 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.466687 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:44:07 crc kubenswrapper[4934]: E1227 07:44:07.466836 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.466901 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:07 crc kubenswrapper[4934]: E1227 07:44:07.467045 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:44:07 crc kubenswrapper[4934]: E1227 07:44:07.467205 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.508182 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.508232 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.508248 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.508280 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.508297 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:07Z","lastTransitionTime":"2025-12-27T07:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.612605 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.612658 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.612674 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.612696 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.612713 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:07Z","lastTransitionTime":"2025-12-27T07:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.715539 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.715587 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.715606 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.715626 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.715643 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:07Z","lastTransitionTime":"2025-12-27T07:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.818941 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.819032 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.819055 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.819080 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.819129 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:07Z","lastTransitionTime":"2025-12-27T07:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.921898 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.921949 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.921966 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.921988 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:07 crc kubenswrapper[4934]: I1227 07:44:07.922008 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:07Z","lastTransitionTime":"2025-12-27T07:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.025668 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.025721 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.025741 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.025764 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.025784 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:08Z","lastTransitionTime":"2025-12-27T07:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.128198 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.128301 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.128322 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.128346 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.128362 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:08Z","lastTransitionTime":"2025-12-27T07:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.230777 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.231193 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.231452 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.231619 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.231745 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:08Z","lastTransitionTime":"2025-12-27T07:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.334875 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.334932 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.334949 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.334973 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.334990 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:08Z","lastTransitionTime":"2025-12-27T07:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.437799 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.437869 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.437880 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.437897 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.437908 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:08Z","lastTransitionTime":"2025-12-27T07:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.467388 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:44:08 crc kubenswrapper[4934]: E1227 07:44:08.467557 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.540061 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.540149 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.540168 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.540195 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.540212 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:08Z","lastTransitionTime":"2025-12-27T07:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.642967 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.643027 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.643043 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.643071 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.643141 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:08Z","lastTransitionTime":"2025-12-27T07:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.745906 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.745973 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.745999 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.746029 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.746052 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:08Z","lastTransitionTime":"2025-12-27T07:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.849491 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.849599 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.849616 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.849645 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.849667 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:08Z","lastTransitionTime":"2025-12-27T07:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.952326 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.952380 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.952396 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.952418 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:08 crc kubenswrapper[4934]: I1227 07:44:08.952436 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:08Z","lastTransitionTime":"2025-12-27T07:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.066476 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.066561 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.066592 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.066621 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.066638 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:09Z","lastTransitionTime":"2025-12-27T07:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.170832 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.170884 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.170901 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.170925 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.170942 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:09Z","lastTransitionTime":"2025-12-27T07:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.274262 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.274343 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.274359 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.274381 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.274401 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:09Z","lastTransitionTime":"2025-12-27T07:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.377641 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.377703 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.377725 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.377754 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.377776 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:09Z","lastTransitionTime":"2025-12-27T07:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.467709 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:09 crc kubenswrapper[4934]: E1227 07:44:09.467873 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.468230 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:09 crc kubenswrapper[4934]: E1227 07:44:09.468336 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.468475 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:44:09 crc kubenswrapper[4934]: E1227 07:44:09.468610 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.480168 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.480228 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.480246 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.480269 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.480286 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:09Z","lastTransitionTime":"2025-12-27T07:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.522910 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=26.52288002 podStartE2EDuration="26.52288002s" podCreationTimestamp="2025-12-27 07:43:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:09.520980778 +0000 UTC m=+110.341421412" watchObservedRunningTime="2025-12-27 07:44:09.52288002 +0000 UTC m=+110.343320644" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.582309 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.582385 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.582402 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.582430 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.582449 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:09Z","lastTransitionTime":"2025-12-27T07:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.640376 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g2dnl" podStartSLOduration=91.640349384 podStartE2EDuration="1m31.640349384s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:09.623210391 +0000 UTC m=+110.443651055" watchObservedRunningTime="2025-12-27 07:44:09.640349384 +0000 UTC m=+110.460789988" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.667413 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=92.667387921 podStartE2EDuration="1m32.667387921s" podCreationTimestamp="2025-12-27 07:42:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:09.667045581 +0000 UTC m=+110.487486185" watchObservedRunningTime="2025-12-27 07:44:09.667387921 +0000 UTC m=+110.487828545" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.667990 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=23.667982767 podStartE2EDuration="23.667982767s" podCreationTimestamp="2025-12-27 07:43:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:09.644410696 +0000 UTC m=+110.464851300" watchObservedRunningTime="2025-12-27 07:44:09.667982767 +0000 UTC m=+110.488423401" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.685770 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.685831 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.685853 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.685877 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.685894 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:09Z","lastTransitionTime":"2025-12-27T07:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.710479 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-hqfw4" podStartSLOduration=91.71045367 podStartE2EDuration="1m31.71045367s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:09.709049421 +0000 UTC m=+110.529490045" watchObservedRunningTime="2025-12-27 07:44:09.71045367 +0000 UTC m=+110.530894304" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.733429 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=86.733397824 podStartE2EDuration="1m26.733397824s" podCreationTimestamp="2025-12-27 07:42:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:09.731052339 +0000 UTC m=+110.551492963" watchObservedRunningTime="2025-12-27 07:44:09.733397824 +0000 UTC m=+110.553838468" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.788180 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.788489 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.788616 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.788720 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.788809 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:09Z","lastTransitionTime":"2025-12-27T07:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.800131 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podStartSLOduration=91.800111236 podStartE2EDuration="1m31.800111236s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:09.799377096 +0000 UTC m=+110.619817730" watchObservedRunningTime="2025-12-27 07:44:09.800111236 +0000 UTC m=+110.620551860" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.800909 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-fwcxz" podStartSLOduration=91.800902858 podStartE2EDuration="1m31.800902858s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:09.783062765 +0000 UTC m=+110.603503369" watchObservedRunningTime="2025-12-27 07:44:09.800902858 +0000 UTC m=+110.621343462" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.839737 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-jzv2s" podStartSLOduration=91.83971586 podStartE2EDuration="1m31.83971586s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:09.822977398 +0000 UTC m=+110.643418022" watchObservedRunningTime="2025-12-27 07:44:09.83971586 +0000 UTC m=+110.660156494" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.865063 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=55.865038079 podStartE2EDuration="55.865038079s" podCreationTimestamp="2025-12-27 07:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:09.844681787 +0000 UTC m=+110.665122391" watchObservedRunningTime="2025-12-27 07:44:09.865038079 +0000 UTC m=+110.685478713" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.891432 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.891520 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.891540 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.891568 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.891586 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:09Z","lastTransitionTime":"2025-12-27T07:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.901372 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-dfj4x" podStartSLOduration=91.901346562 podStartE2EDuration="1m31.901346562s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:09.888667712 +0000 UTC m=+110.709108336" watchObservedRunningTime="2025-12-27 07:44:09.901346562 +0000 UTC m=+110.721787196" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.994275 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.994343 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.994364 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.994388 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:09 crc kubenswrapper[4934]: I1227 07:44:09.994407 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:09Z","lastTransitionTime":"2025-12-27T07:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.096609 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.096658 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.096669 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.096685 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.096695 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:10Z","lastTransitionTime":"2025-12-27T07:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.198952 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.199000 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.199010 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.199028 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.199041 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:10Z","lastTransitionTime":"2025-12-27T07:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.302368 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.302420 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.302429 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.302442 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.302452 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:10Z","lastTransitionTime":"2025-12-27T07:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.404808 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.404882 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.404894 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.404911 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.404923 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:10Z","lastTransitionTime":"2025-12-27T07:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.466387 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:44:10 crc kubenswrapper[4934]: E1227 07:44:10.466514 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.514714 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.514785 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.514806 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.514839 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.514861 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:10Z","lastTransitionTime":"2025-12-27T07:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.616867 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.616896 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.616929 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.616945 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.616955 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:10Z","lastTransitionTime":"2025-12-27T07:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.720036 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.720162 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.720181 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.720206 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.720225 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:10Z","lastTransitionTime":"2025-12-27T07:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.823326 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.823381 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.823399 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.823421 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.823439 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:10Z","lastTransitionTime":"2025-12-27T07:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.926731 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.926786 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.926804 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.926826 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:10 crc kubenswrapper[4934]: I1227 07:44:10.926843 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:10Z","lastTransitionTime":"2025-12-27T07:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.029671 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.029729 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.029746 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.029772 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.029795 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:11Z","lastTransitionTime":"2025-12-27T07:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.132718 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.132774 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.132792 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.132816 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.132837 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:11Z","lastTransitionTime":"2025-12-27T07:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.235706 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.235763 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.235781 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.235809 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.235829 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:11Z","lastTransitionTime":"2025-12-27T07:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.338763 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.338852 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.338870 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.338893 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.338912 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:11Z","lastTransitionTime":"2025-12-27T07:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.442156 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.442235 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.442258 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.442288 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.442309 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:11Z","lastTransitionTime":"2025-12-27T07:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.466796 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.466814 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:11 crc kubenswrapper[4934]: E1227 07:44:11.466999 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.467062 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:11 crc kubenswrapper[4934]: E1227 07:44:11.467219 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:44:11 crc kubenswrapper[4934]: E1227 07:44:11.467454 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.545355 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.545414 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.545432 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.545458 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.545475 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:11Z","lastTransitionTime":"2025-12-27T07:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.648317 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.648386 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.648410 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.648435 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.648452 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:11Z","lastTransitionTime":"2025-12-27T07:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.751967 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.752075 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.752124 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.752149 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.752168 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:11Z","lastTransitionTime":"2025-12-27T07:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.854928 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.854983 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.855000 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.855028 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.855046 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:11Z","lastTransitionTime":"2025-12-27T07:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.958317 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.958375 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.958390 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.958410 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:11 crc kubenswrapper[4934]: I1227 07:44:11.958425 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:11Z","lastTransitionTime":"2025-12-27T07:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.060709 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.060761 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.060769 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.060784 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.060795 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:12Z","lastTransitionTime":"2025-12-27T07:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.113045 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jzv2s_1bdd4b78-f615-4771-9448-8faa03dbc4f5/kube-multus/1.log" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.113833 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jzv2s_1bdd4b78-f615-4771-9448-8faa03dbc4f5/kube-multus/0.log" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.113907 4934 generic.go:334] "Generic (PLEG): container finished" podID="1bdd4b78-f615-4771-9448-8faa03dbc4f5" containerID="dd93f8a550f1648117195c4caac3c1e150c84a1c2dba90ae1ba1a4ecc1bd284b" exitCode=1 Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.113948 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jzv2s" event={"ID":"1bdd4b78-f615-4771-9448-8faa03dbc4f5","Type":"ContainerDied","Data":"dd93f8a550f1648117195c4caac3c1e150c84a1c2dba90ae1ba1a4ecc1bd284b"} Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.113992 4934 scope.go:117] "RemoveContainer" containerID="a8e52f0dc2c24df10628df4a52757af2183a1815c1b388166c468a740adc8c44" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.114715 4934 scope.go:117] "RemoveContainer" containerID="dd93f8a550f1648117195c4caac3c1e150c84a1c2dba90ae1ba1a4ecc1bd284b" Dec 27 07:44:12 crc kubenswrapper[4934]: E1227 07:44:12.115211 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-jzv2s_openshift-multus(1bdd4b78-f615-4771-9448-8faa03dbc4f5)\"" pod="openshift-multus/multus-jzv2s" podUID="1bdd4b78-f615-4771-9448-8faa03dbc4f5" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.162909 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.162965 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.162979 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.163002 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.163019 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:12Z","lastTransitionTime":"2025-12-27T07:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.265885 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.265936 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.265954 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.265978 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.265994 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:12Z","lastTransitionTime":"2025-12-27T07:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.369745 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.370155 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.370419 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.370624 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.370764 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:12Z","lastTransitionTime":"2025-12-27T07:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.466941 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:44:12 crc kubenswrapper[4934]: E1227 07:44:12.467157 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.473192 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.473560 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.473746 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.473953 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.474188 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:12Z","lastTransitionTime":"2025-12-27T07:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.578060 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.578171 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.578193 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.578219 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.578239 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:12Z","lastTransitionTime":"2025-12-27T07:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.681798 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.681869 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.681894 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.681930 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.681956 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:12Z","lastTransitionTime":"2025-12-27T07:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.785334 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.785400 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.785420 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.785444 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.785463 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:12Z","lastTransitionTime":"2025-12-27T07:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.888306 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.888357 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.888374 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.888395 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.888413 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:12Z","lastTransitionTime":"2025-12-27T07:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.990860 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.990910 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.990927 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.990951 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:12 crc kubenswrapper[4934]: I1227 07:44:12.990968 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:12Z","lastTransitionTime":"2025-12-27T07:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.094050 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.094461 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.094648 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.094874 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.095135 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:13Z","lastTransitionTime":"2025-12-27T07:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.119303 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jzv2s_1bdd4b78-f615-4771-9448-8faa03dbc4f5/kube-multus/1.log" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.197614 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.198286 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.198331 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.198351 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.198365 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:13Z","lastTransitionTime":"2025-12-27T07:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.301532 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.301612 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.301638 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.301672 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.301692 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:13Z","lastTransitionTime":"2025-12-27T07:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.404295 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.404354 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.404371 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.404394 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.404411 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:13Z","lastTransitionTime":"2025-12-27T07:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.467344 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.467355 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:44:13 crc kubenswrapper[4934]: E1227 07:44:13.467545 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:44:13 crc kubenswrapper[4934]: E1227 07:44:13.467704 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.467970 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:13 crc kubenswrapper[4934]: E1227 07:44:13.468271 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.507529 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.507624 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.507643 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.507700 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.507720 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:13Z","lastTransitionTime":"2025-12-27T07:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.611318 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.611738 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.611926 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.612208 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.612581 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:13Z","lastTransitionTime":"2025-12-27T07:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.715414 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.715664 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.715752 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.715839 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.715990 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:13Z","lastTransitionTime":"2025-12-27T07:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.752879 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.753247 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.753430 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.753647 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.753848 4934 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-27T07:44:13Z","lastTransitionTime":"2025-12-27T07:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.823415 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-6q52c"] Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.824210 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6q52c" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.827920 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.828440 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.828033 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.828166 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.883026 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/144b3757-23f0-4dc9-8a29-b51a42188c40-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-6q52c\" (UID: \"144b3757-23f0-4dc9-8a29-b51a42188c40\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6q52c" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.883133 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/144b3757-23f0-4dc9-8a29-b51a42188c40-service-ca\") pod \"cluster-version-operator-5c965bbfc6-6q52c\" (UID: \"144b3757-23f0-4dc9-8a29-b51a42188c40\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6q52c" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.883330 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/144b3757-23f0-4dc9-8a29-b51a42188c40-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-6q52c\" (UID: \"144b3757-23f0-4dc9-8a29-b51a42188c40\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6q52c" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.883393 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/144b3757-23f0-4dc9-8a29-b51a42188c40-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-6q52c\" (UID: \"144b3757-23f0-4dc9-8a29-b51a42188c40\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6q52c" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.883486 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/144b3757-23f0-4dc9-8a29-b51a42188c40-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-6q52c\" (UID: \"144b3757-23f0-4dc9-8a29-b51a42188c40\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6q52c" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.984658 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/144b3757-23f0-4dc9-8a29-b51a42188c40-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-6q52c\" (UID: \"144b3757-23f0-4dc9-8a29-b51a42188c40\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6q52c" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.984733 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/144b3757-23f0-4dc9-8a29-b51a42188c40-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-6q52c\" (UID: \"144b3757-23f0-4dc9-8a29-b51a42188c40\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6q52c" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.984827 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/144b3757-23f0-4dc9-8a29-b51a42188c40-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-6q52c\" (UID: \"144b3757-23f0-4dc9-8a29-b51a42188c40\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6q52c" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.984902 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/144b3757-23f0-4dc9-8a29-b51a42188c40-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-6q52c\" (UID: \"144b3757-23f0-4dc9-8a29-b51a42188c40\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6q52c" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.984956 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/144b3757-23f0-4dc9-8a29-b51a42188c40-service-ca\") pod \"cluster-version-operator-5c965bbfc6-6q52c\" (UID: \"144b3757-23f0-4dc9-8a29-b51a42188c40\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6q52c" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.985403 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/144b3757-23f0-4dc9-8a29-b51a42188c40-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-6q52c\" (UID: \"144b3757-23f0-4dc9-8a29-b51a42188c40\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6q52c" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.985432 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/144b3757-23f0-4dc9-8a29-b51a42188c40-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-6q52c\" (UID: \"144b3757-23f0-4dc9-8a29-b51a42188c40\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6q52c" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.986949 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/144b3757-23f0-4dc9-8a29-b51a42188c40-service-ca\") pod \"cluster-version-operator-5c965bbfc6-6q52c\" (UID: \"144b3757-23f0-4dc9-8a29-b51a42188c40\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6q52c" Dec 27 07:44:13 crc kubenswrapper[4934]: I1227 07:44:13.996654 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/144b3757-23f0-4dc9-8a29-b51a42188c40-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-6q52c\" (UID: \"144b3757-23f0-4dc9-8a29-b51a42188c40\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6q52c" Dec 27 07:44:14 crc kubenswrapper[4934]: I1227 07:44:14.012577 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/144b3757-23f0-4dc9-8a29-b51a42188c40-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-6q52c\" (UID: \"144b3757-23f0-4dc9-8a29-b51a42188c40\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6q52c" Dec 27 07:44:14 crc kubenswrapper[4934]: I1227 07:44:14.152804 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6q52c" Dec 27 07:44:14 crc kubenswrapper[4934]: I1227 07:44:14.466973 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:44:14 crc kubenswrapper[4934]: E1227 07:44:14.467961 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:44:14 crc kubenswrapper[4934]: I1227 07:44:14.468138 4934 scope.go:117] "RemoveContainer" containerID="0133bfa46253ea9cf0c58a1951b5b012f93f5e8e1ae0ce6bff2341e6d94366c4" Dec 27 07:44:14 crc kubenswrapper[4934]: E1227 07:44:14.468874 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8vzdl_openshift-ovn-kubernetes(1c4780a7-56b7-4772-af66-045a03b31a71)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" Dec 27 07:44:15 crc kubenswrapper[4934]: I1227 07:44:15.130185 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6q52c" event={"ID":"144b3757-23f0-4dc9-8a29-b51a42188c40","Type":"ContainerStarted","Data":"da1f235e3a32e6d621e0e45c8ae2849a4c482d6a6f97404d7394d7ce11aea71d"} Dec 27 07:44:15 crc kubenswrapper[4934]: I1227 07:44:15.130801 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6q52c" event={"ID":"144b3757-23f0-4dc9-8a29-b51a42188c40","Type":"ContainerStarted","Data":"6db3282474431f6e1d9fa8de40a5236a3a99e451af977744b5bf5fd53104714b"} Dec 27 07:44:15 crc kubenswrapper[4934]: I1227 07:44:15.151518 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-6q52c" podStartSLOduration=97.151489252 podStartE2EDuration="1m37.151489252s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:15.149493027 +0000 UTC m=+115.969933631" watchObservedRunningTime="2025-12-27 07:44:15.151489252 +0000 UTC m=+115.971929886" Dec 27 07:44:15 crc kubenswrapper[4934]: I1227 07:44:15.466858 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:15 crc kubenswrapper[4934]: I1227 07:44:15.466996 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:44:15 crc kubenswrapper[4934]: E1227 07:44:15.467054 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:44:15 crc kubenswrapper[4934]: E1227 07:44:15.467273 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:44:15 crc kubenswrapper[4934]: I1227 07:44:15.467371 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:15 crc kubenswrapper[4934]: E1227 07:44:15.467510 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:44:16 crc kubenswrapper[4934]: I1227 07:44:16.467207 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:44:16 crc kubenswrapper[4934]: E1227 07:44:16.468367 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:44:17 crc kubenswrapper[4934]: I1227 07:44:17.466440 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:17 crc kubenswrapper[4934]: I1227 07:44:17.466490 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:17 crc kubenswrapper[4934]: I1227 07:44:17.466554 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:44:17 crc kubenswrapper[4934]: E1227 07:44:17.466664 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:44:17 crc kubenswrapper[4934]: E1227 07:44:17.466856 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:44:17 crc kubenswrapper[4934]: E1227 07:44:17.466969 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:44:18 crc kubenswrapper[4934]: I1227 07:44:18.466549 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:44:18 crc kubenswrapper[4934]: E1227 07:44:18.466696 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:44:19 crc kubenswrapper[4934]: I1227 07:44:19.466858 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:19 crc kubenswrapper[4934]: I1227 07:44:19.466932 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:19 crc kubenswrapper[4934]: I1227 07:44:19.466978 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:44:19 crc kubenswrapper[4934]: E1227 07:44:19.468390 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:44:19 crc kubenswrapper[4934]: E1227 07:44:19.468056 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:44:19 crc kubenswrapper[4934]: E1227 07:44:19.468814 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:44:19 crc kubenswrapper[4934]: E1227 07:44:19.489860 4934 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 27 07:44:19 crc kubenswrapper[4934]: E1227 07:44:19.587498 4934 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 27 07:44:20 crc kubenswrapper[4934]: I1227 07:44:20.467131 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:44:20 crc kubenswrapper[4934]: E1227 07:44:20.467314 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:44:21 crc kubenswrapper[4934]: I1227 07:44:21.466950 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:44:21 crc kubenswrapper[4934]: I1227 07:44:21.467060 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:21 crc kubenswrapper[4934]: E1227 07:44:21.467558 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:44:21 crc kubenswrapper[4934]: E1227 07:44:21.467423 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:44:21 crc kubenswrapper[4934]: I1227 07:44:21.467122 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:21 crc kubenswrapper[4934]: E1227 07:44:21.467625 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:44:22 crc kubenswrapper[4934]: I1227 07:44:22.466563 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:44:22 crc kubenswrapper[4934]: E1227 07:44:22.466756 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:44:23 crc kubenswrapper[4934]: I1227 07:44:23.467142 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:44:23 crc kubenswrapper[4934]: I1227 07:44:23.467230 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:23 crc kubenswrapper[4934]: I1227 07:44:23.467230 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:23 crc kubenswrapper[4934]: E1227 07:44:23.467368 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:44:23 crc kubenswrapper[4934]: E1227 07:44:23.467533 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:44:23 crc kubenswrapper[4934]: E1227 07:44:23.467734 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:44:24 crc kubenswrapper[4934]: I1227 07:44:24.466739 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:44:24 crc kubenswrapper[4934]: E1227 07:44:24.467051 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:44:24 crc kubenswrapper[4934]: I1227 07:44:24.467624 4934 scope.go:117] "RemoveContainer" containerID="dd93f8a550f1648117195c4caac3c1e150c84a1c2dba90ae1ba1a4ecc1bd284b" Dec 27 07:44:24 crc kubenswrapper[4934]: E1227 07:44:24.589370 4934 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 27 07:44:25 crc kubenswrapper[4934]: I1227 07:44:25.176494 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jzv2s_1bdd4b78-f615-4771-9448-8faa03dbc4f5/kube-multus/1.log" Dec 27 07:44:25 crc kubenswrapper[4934]: I1227 07:44:25.176573 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jzv2s" event={"ID":"1bdd4b78-f615-4771-9448-8faa03dbc4f5","Type":"ContainerStarted","Data":"b3182ba625b9908dd1c4de2aa6395b330f542942567d562205e6ee11080b342a"} Dec 27 07:44:25 crc kubenswrapper[4934]: I1227 07:44:25.466621 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:44:25 crc kubenswrapper[4934]: I1227 07:44:25.466660 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:25 crc kubenswrapper[4934]: I1227 07:44:25.466656 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:25 crc kubenswrapper[4934]: E1227 07:44:25.466839 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:44:25 crc kubenswrapper[4934]: E1227 07:44:25.466928 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:44:25 crc kubenswrapper[4934]: E1227 07:44:25.467011 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:44:26 crc kubenswrapper[4934]: I1227 07:44:26.466607 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:44:26 crc kubenswrapper[4934]: E1227 07:44:26.466839 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:44:27 crc kubenswrapper[4934]: I1227 07:44:27.467360 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:27 crc kubenswrapper[4934]: E1227 07:44:27.467560 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:44:27 crc kubenswrapper[4934]: I1227 07:44:27.467662 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:27 crc kubenswrapper[4934]: E1227 07:44:27.467916 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:44:27 crc kubenswrapper[4934]: I1227 07:44:27.468017 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:44:27 crc kubenswrapper[4934]: E1227 07:44:27.468327 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:44:28 crc kubenswrapper[4934]: I1227 07:44:28.466889 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:44:28 crc kubenswrapper[4934]: E1227 07:44:28.467186 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:44:29 crc kubenswrapper[4934]: I1227 07:44:29.467225 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:29 crc kubenswrapper[4934]: E1227 07:44:29.469265 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:44:29 crc kubenswrapper[4934]: I1227 07:44:29.469378 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:29 crc kubenswrapper[4934]: E1227 07:44:29.469514 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:44:29 crc kubenswrapper[4934]: I1227 07:44:29.469519 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:44:29 crc kubenswrapper[4934]: E1227 07:44:29.470116 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:44:29 crc kubenswrapper[4934]: I1227 07:44:29.470634 4934 scope.go:117] "RemoveContainer" containerID="0133bfa46253ea9cf0c58a1951b5b012f93f5e8e1ae0ce6bff2341e6d94366c4" Dec 27 07:44:29 crc kubenswrapper[4934]: E1227 07:44:29.590193 4934 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 27 07:44:30 crc kubenswrapper[4934]: I1227 07:44:30.198974 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovnkube-controller/3.log" Dec 27 07:44:30 crc kubenswrapper[4934]: I1227 07:44:30.202747 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerStarted","Data":"478607a19e4255f7af127be496f42f13353eac046569986baf8095c3860f53c9"} Dec 27 07:44:30 crc kubenswrapper[4934]: I1227 07:44:30.203310 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:44:30 crc kubenswrapper[4934]: I1227 07:44:30.235877 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" podStartSLOduration=112.235853133 podStartE2EDuration="1m52.235853133s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:30.233849357 +0000 UTC m=+131.054289951" watchObservedRunningTime="2025-12-27 07:44:30.235853133 +0000 UTC m=+131.056293757" Dec 27 07:44:30 crc kubenswrapper[4934]: I1227 07:44:30.466844 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:44:30 crc kubenswrapper[4934]: E1227 07:44:30.467036 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:44:30 crc kubenswrapper[4934]: I1227 07:44:30.652551 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-jjlqg"] Dec 27 07:44:31 crc kubenswrapper[4934]: I1227 07:44:31.206574 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:44:31 crc kubenswrapper[4934]: E1227 07:44:31.207230 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:44:31 crc kubenswrapper[4934]: I1227 07:44:31.467144 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:31 crc kubenswrapper[4934]: E1227 07:44:31.467390 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:44:31 crc kubenswrapper[4934]: I1227 07:44:31.467420 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:31 crc kubenswrapper[4934]: E1227 07:44:31.467614 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:44:31 crc kubenswrapper[4934]: I1227 07:44:31.467697 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:44:31 crc kubenswrapper[4934]: E1227 07:44:31.467914 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:44:33 crc kubenswrapper[4934]: I1227 07:44:33.467333 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:33 crc kubenswrapper[4934]: I1227 07:44:33.467391 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:44:33 crc kubenswrapper[4934]: I1227 07:44:33.467421 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:33 crc kubenswrapper[4934]: E1227 07:44:33.467566 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 27 07:44:33 crc kubenswrapper[4934]: I1227 07:44:33.467642 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:44:33 crc kubenswrapper[4934]: E1227 07:44:33.467773 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 27 07:44:33 crc kubenswrapper[4934]: E1227 07:44:33.467935 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-jjlqg" podUID="304709d3-05ab-4d27-8722-f94053159a29" Dec 27 07:44:33 crc kubenswrapper[4934]: E1227 07:44:33.468033 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 27 07:44:35 crc kubenswrapper[4934]: I1227 07:44:35.467299 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:35 crc kubenswrapper[4934]: I1227 07:44:35.467421 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:44:35 crc kubenswrapper[4934]: I1227 07:44:35.467535 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:44:35 crc kubenswrapper[4934]: I1227 07:44:35.467740 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:35 crc kubenswrapper[4934]: I1227 07:44:35.475456 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 27 07:44:35 crc kubenswrapper[4934]: I1227 07:44:35.475857 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 27 07:44:35 crc kubenswrapper[4934]: I1227 07:44:35.476240 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 27 07:44:35 crc kubenswrapper[4934]: I1227 07:44:35.477028 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 27 07:44:35 crc kubenswrapper[4934]: I1227 07:44:35.477578 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 27 07:44:35 crc kubenswrapper[4934]: I1227 07:44:35.478135 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 27 07:44:38 crc kubenswrapper[4934]: I1227 07:44:38.278050 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.467237 4934 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.522072 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dwv8c"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.522640 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.523623 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-jckgt"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.525219 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8jpzb"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.525397 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-jckgt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.526543 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8jpzb" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.529386 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.529660 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.530562 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-p8c2x"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.531308 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.533076 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9s2ls"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.533609 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9s2ls" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.534006 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-j5cpz"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.534981 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.535262 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.537137 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.539571 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.540109 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.540276 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.540405 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.540505 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.540529 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.540761 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.540834 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.540972 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.541097 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.541166 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.541196 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.541324 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.541341 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.541389 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.541724 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.541824 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.541884 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-8msj4"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.542178 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.542264 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.542341 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.542569 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8msj4" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.543335 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.543364 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.543918 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.544648 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hn7lm"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.545122 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hn7lm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.545475 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-whhqx"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.545960 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-whhqx" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.547886 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.549218 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.549980 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ln99b"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.566939 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22jvt"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.567627 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hh662"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.568702 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zcl7d"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.570363 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.570782 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22jvt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.571733 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.573688 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.577110 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.583946 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.584799 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-nr4vm"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.585213 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.585756 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zcl7d" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.585880 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.586066 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.586486 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.587141 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hh662" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.588518 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.588830 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.589126 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.589201 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.589243 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-2zfd9"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.590509 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-z76rm"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.590964 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.591322 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.591408 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ph956"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.591541 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.591658 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2zfd9" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.591949 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-z76rm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.592011 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8npds"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.592402 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ph956" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.592598 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.592764 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.593454 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.593637 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.594509 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.596218 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gdpgt"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.597609 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.597782 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.597893 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.598092 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.598245 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.598393 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.598480 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.598519 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.598550 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.598672 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.598901 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.599013 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.599114 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.599346 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.599645 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gdpgt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.600767 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.600882 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.601515 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.601584 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.602137 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.602331 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.602552 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.608868 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ndhjw"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.608930 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.609707 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ndhjw" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.610627 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wvjnl"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.611140 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8s8bp"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.611751 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.611865 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mn55p"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.612033 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wvjnl" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.612961 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.613292 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.613570 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn55p" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.613596 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.613846 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.613859 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.614056 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.629902 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.631873 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.632477 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.639012 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.640260 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.640528 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.640831 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-w4pmg"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.640907 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.641320 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.641566 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.641783 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6bmfr"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.642444 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.642702 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-w4pmg" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.642765 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.643184 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.643431 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.643654 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.643663 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.643699 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.643854 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.647774 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.647902 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.647986 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.648046 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.648138 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.648521 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.648659 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.648774 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.648865 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.648917 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bmfr" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.649042 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.649068 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.649219 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.650389 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655197 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655266 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-console-oauth-config\") pod \"console-f9d7485db-p8c2x\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655307 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655338 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6342508-1473-49c4-a78f-b4e6ed9c68b0-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-9s2ls\" (UID: \"d6342508-1473-49c4-a78f-b4e6ed9c68b0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9s2ls" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655371 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-audit-policies\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655391 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-service-ca\") pod \"console-f9d7485db-p8c2x\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655408 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655439 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q97bs\" (UniqueName: \"kubernetes.io/projected/753c599b-a401-481e-a5b3-e2dd6ed2c4e7-kube-api-access-q97bs\") pod \"console-operator-58897d9998-jckgt\" (UID: \"753c599b-a401-481e-a5b3-e2dd6ed2c4e7\") " pod="openshift-console-operator/console-operator-58897d9998-jckgt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655456 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/ba604131-ba14-4754-8c5e-3e4fdccdfe64-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8jpzb\" (UID: \"ba604131-ba14-4754-8c5e-3e4fdccdfe64\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8jpzb" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655473 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-trusted-ca-bundle\") pod \"console-f9d7485db-p8c2x\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655489 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/753c599b-a401-481e-a5b3-e2dd6ed2c4e7-trusted-ca\") pod \"console-operator-58897d9998-jckgt\" (UID: \"753c599b-a401-481e-a5b3-e2dd6ed2c4e7\") " pod="openshift-console-operator/console-operator-58897d9998-jckgt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655505 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ebeee10-d90f-4672-8bfa-5f10fa929d31-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hn7lm\" (UID: \"7ebeee10-d90f-4672-8bfa-5f10fa929d31\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hn7lm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655524 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ebeee10-d90f-4672-8bfa-5f10fa929d31-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hn7lm\" (UID: \"7ebeee10-d90f-4672-8bfa-5f10fa929d31\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hn7lm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655544 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpk5n\" (UniqueName: \"kubernetes.io/projected/c2f61301-d046-4651-b76c-9939b18fb792-kube-api-access-jpk5n\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655566 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn7p8\" (UniqueName: \"kubernetes.io/projected/7ebeee10-d90f-4672-8bfa-5f10fa929d31-kube-api-access-cn7p8\") pod \"openshift-controller-manager-operator-756b6f6bc6-hn7lm\" (UID: \"7ebeee10-d90f-4672-8bfa-5f10fa929d31\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hn7lm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655582 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-oauth-serving-cert\") pod \"console-f9d7485db-p8c2x\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655596 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/959b5ca2-d235-4457-98ec-fe077dfcdc6f-serving-cert\") pod \"route-controller-manager-6576b87f9c-r5zxc\" (UID: \"959b5ca2-d235-4457-98ec-fe077dfcdc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655616 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/901df0a0-1dae-41aa-8d29-0981524d9312-serving-cert\") pod \"authentication-operator-69f744f599-dwv8c\" (UID: \"901df0a0-1dae-41aa-8d29-0981524d9312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655632 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78dm8\" (UniqueName: \"kubernetes.io/projected/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-kube-api-access-78dm8\") pod \"console-f9d7485db-p8c2x\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655649 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwq7v\" (UniqueName: \"kubernetes.io/projected/d6342508-1473-49c4-a78f-b4e6ed9c68b0-kube-api-access-mwq7v\") pod \"openshift-apiserver-operator-796bbdcf4f-9s2ls\" (UID: \"d6342508-1473-49c4-a78f-b4e6ed9c68b0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9s2ls" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655662 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655676 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/959b5ca2-d235-4457-98ec-fe077dfcdc6f-config\") pod \"route-controller-manager-6576b87f9c-r5zxc\" (UID: \"959b5ca2-d235-4457-98ec-fe077dfcdc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655689 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcpfs\" (UniqueName: \"kubernetes.io/projected/ba604131-ba14-4754-8c5e-3e4fdccdfe64-kube-api-access-zcpfs\") pod \"cluster-samples-operator-665b6dd947-8jpzb\" (UID: \"ba604131-ba14-4754-8c5e-3e4fdccdfe64\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8jpzb" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655712 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655727 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/901df0a0-1dae-41aa-8d29-0981524d9312-config\") pod \"authentication-operator-69f744f599-dwv8c\" (UID: \"901df0a0-1dae-41aa-8d29-0981524d9312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655745 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22f69\" (UniqueName: \"kubernetes.io/projected/901df0a0-1dae-41aa-8d29-0981524d9312-kube-api-access-22f69\") pod \"authentication-operator-69f744f599-dwv8c\" (UID: \"901df0a0-1dae-41aa-8d29-0981524d9312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655764 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c2f61301-d046-4651-b76c-9939b18fb792-audit-dir\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655797 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655815 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/753c599b-a401-481e-a5b3-e2dd6ed2c4e7-config\") pod \"console-operator-58897d9998-jckgt\" (UID: \"753c599b-a401-481e-a5b3-e2dd6ed2c4e7\") " pod="openshift-console-operator/console-operator-58897d9998-jckgt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655836 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655855 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwk6g\" (UniqueName: \"kubernetes.io/projected/959b5ca2-d235-4457-98ec-fe077dfcdc6f-kube-api-access-wwk6g\") pod \"route-controller-manager-6576b87f9c-r5zxc\" (UID: \"959b5ca2-d235-4457-98ec-fe077dfcdc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655869 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655888 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655901 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/901df0a0-1dae-41aa-8d29-0981524d9312-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dwv8c\" (UID: \"901df0a0-1dae-41aa-8d29-0981524d9312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655917 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/901df0a0-1dae-41aa-8d29-0981524d9312-service-ca-bundle\") pod \"authentication-operator-69f744f599-dwv8c\" (UID: \"901df0a0-1dae-41aa-8d29-0981524d9312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655966 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655981 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-console-serving-cert\") pod \"console-f9d7485db-p8c2x\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.655995 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkkfs\" (UniqueName: \"kubernetes.io/projected/a0d34185-bf08-456c-a5fc-3f0852d35a91-kube-api-access-lkkfs\") pod \"downloads-7954f5f757-8msj4\" (UID: \"a0d34185-bf08-456c-a5fc-3f0852d35a91\") " pod="openshift-console/downloads-7954f5f757-8msj4" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.656009 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.656024 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6342508-1473-49c4-a78f-b4e6ed9c68b0-config\") pod \"openshift-apiserver-operator-796bbdcf4f-9s2ls\" (UID: \"d6342508-1473-49c4-a78f-b4e6ed9c68b0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9s2ls" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.656039 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/753c599b-a401-481e-a5b3-e2dd6ed2c4e7-serving-cert\") pod \"console-operator-58897d9998-jckgt\" (UID: \"753c599b-a401-481e-a5b3-e2dd6ed2c4e7\") " pod="openshift-console-operator/console-operator-58897d9998-jckgt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.656053 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/959b5ca2-d235-4457-98ec-fe077dfcdc6f-client-ca\") pod \"route-controller-manager-6576b87f9c-r5zxc\" (UID: \"959b5ca2-d235-4457-98ec-fe077dfcdc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.656070 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-console-config\") pod \"console-f9d7485db-p8c2x\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.656190 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.656528 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.656604 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.656667 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.656678 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.657668 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.658476 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xxrp6"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.659134 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.660052 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-c2tqv"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.660769 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-c2tqv" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.660953 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dwv8c"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.661903 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.662503 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-hrm6w"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.663275 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hrm6w" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.664219 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-spr46"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.664637 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-spr46" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.665585 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.666094 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.666406 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zchqn"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.667473 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zchqn" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.668272 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.668613 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.668720 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.670762 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-cpts5"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.671515 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-cpts5" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.672769 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9s2ls"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.673231 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8jpzb"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.677013 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.677915 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.684264 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8msj4"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.687241 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-j5cpz"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.689387 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.693842 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22jvt"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.697634 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-jckgt"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.698843 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gdpgt"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.700100 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8npds"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.701097 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ln99b"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.703330 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zcl7d"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.704504 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-p8c2x"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.704627 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.705605 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-whhqx"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.708766 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.711987 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wvjnl"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.713378 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mn55p"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.714833 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.716245 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-7tf92"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.717056 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-7tf92" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.719688 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-z76rm"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.725703 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.729255 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6bmfr"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.730067 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hn7lm"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.731465 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.733097 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xxrp6"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.734844 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ph956"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.736415 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8s8bp"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.737589 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hh662"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.739285 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-spr46"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.741238 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-c2tqv"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.743028 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.744186 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.745282 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.745335 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-hrm6w"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.746795 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-nr4vm"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.748332 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ndhjw"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.750042 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-5s9zf"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.751282 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-md2nq"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.751541 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.752021 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-md2nq" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.752469 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-w4pmg"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.753532 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-5s9zf"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.754851 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-md2nq"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.757429 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/753c599b-a401-481e-a5b3-e2dd6ed2c4e7-trusted-ca\") pod \"console-operator-58897d9998-jckgt\" (UID: \"753c599b-a401-481e-a5b3-e2dd6ed2c4e7\") " pod="openshift-console-operator/console-operator-58897d9998-jckgt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.757468 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ebeee10-d90f-4672-8bfa-5f10fa929d31-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hn7lm\" (UID: \"7ebeee10-d90f-4672-8bfa-5f10fa929d31\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hn7lm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.757499 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2ktb\" (UniqueName: \"kubernetes.io/projected/bd2142a6-e9cb-432d-adf9-61b5d3f3c5b1-kube-api-access-m2ktb\") pod \"service-ca-9c57cc56f-w4pmg\" (UID: \"bd2142a6-e9cb-432d-adf9-61b5d3f3c5b1\") " pod="openshift-service-ca/service-ca-9c57cc56f-w4pmg" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.757522 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/145ee02a-c7aa-4990-a288-e6f69f8ab030-encryption-config\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.757547 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-trusted-ca-bundle\") pod \"console-f9d7485db-p8c2x\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.757569 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ebeee10-d90f-4672-8bfa-5f10fa929d31-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hn7lm\" (UID: \"7ebeee10-d90f-4672-8bfa-5f10fa929d31\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hn7lm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.757591 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/bd2142a6-e9cb-432d-adf9-61b5d3f3c5b1-signing-cabundle\") pod \"service-ca-9c57cc56f-w4pmg\" (UID: \"bd2142a6-e9cb-432d-adf9-61b5d3f3c5b1\") " pod="openshift-service-ca/service-ca-9c57cc56f-w4pmg" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.757613 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxj8b\" (UniqueName: \"kubernetes.io/projected/46527d26-eb38-494f-85f7-f5f56b8dfada-kube-api-access-qxj8b\") pod \"machine-api-operator-5694c8668f-whhqx\" (UID: \"46527d26-eb38-494f-85f7-f5f56b8dfada\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-whhqx" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.757880 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.758848 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b2a55db0-a3bd-413f-a52f-75b36654377c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mn55p\" (UID: \"b2a55db0-a3bd-413f-a52f-75b36654377c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn55p" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.758906 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24q7q\" (UniqueName: \"kubernetes.io/projected/af023fee-eae1-427a-8d70-487588422316-kube-api-access-24q7q\") pod \"catalog-operator-68c6474976-5cc9p\" (UID: \"af023fee-eae1-427a-8d70-487588422316\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.758944 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpk5n\" (UniqueName: \"kubernetes.io/projected/c2f61301-d046-4651-b76c-9939b18fb792-kube-api-access-jpk5n\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.758974 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn7p8\" (UniqueName: \"kubernetes.io/projected/7ebeee10-d90f-4672-8bfa-5f10fa929d31-kube-api-access-cn7p8\") pod \"openshift-controller-manager-operator-756b6f6bc6-hn7lm\" (UID: \"7ebeee10-d90f-4672-8bfa-5f10fa929d31\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hn7lm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.759016 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzlhf\" (UniqueName: \"kubernetes.io/projected/0b1dac22-8539-4f61-ac28-2f420c845a2a-kube-api-access-hzlhf\") pod \"kube-storage-version-migrator-operator-b67b599dd-hh662\" (UID: \"0b1dac22-8539-4f61-ac28-2f420c845a2a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hh662" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.759050 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d812eb7a-ce30-4759-a558-fff3738068b6-etcd-client\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.759075 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d812eb7a-ce30-4759-a558-fff3738068b6-serving-cert\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.759124 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d812eb7a-ce30-4759-a558-fff3738068b6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.759162 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/901df0a0-1dae-41aa-8d29-0981524d9312-serving-cert\") pod \"authentication-operator-69f744f599-dwv8c\" (UID: \"901df0a0-1dae-41aa-8d29-0981524d9312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.759188 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1e4154b-2230-4499-b645-4903ea9593f1-config\") pod \"kube-apiserver-operator-766d6c64bb-zcl7d\" (UID: \"a1e4154b-2230-4499-b645-4903ea9593f1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zcl7d" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.759215 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-oauth-serving-cert\") pod \"console-f9d7485db-p8c2x\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.759339 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/959b5ca2-d235-4457-98ec-fe077dfcdc6f-serving-cert\") pod \"route-controller-manager-6576b87f9c-r5zxc\" (UID: \"959b5ca2-d235-4457-98ec-fe077dfcdc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.759369 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwq7v\" (UniqueName: \"kubernetes.io/projected/d6342508-1473-49c4-a78f-b4e6ed9c68b0-kube-api-access-mwq7v\") pod \"openshift-apiserver-operator-796bbdcf4f-9s2ls\" (UID: \"d6342508-1473-49c4-a78f-b4e6ed9c68b0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9s2ls" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.759723 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.759749 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/959b5ca2-d235-4457-98ec-fe077dfcdc6f-config\") pod \"route-controller-manager-6576b87f9c-r5zxc\" (UID: \"959b5ca2-d235-4457-98ec-fe077dfcdc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.759779 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcpfs\" (UniqueName: \"kubernetes.io/projected/ba604131-ba14-4754-8c5e-3e4fdccdfe64-kube-api-access-zcpfs\") pod \"cluster-samples-operator-665b6dd947-8jpzb\" (UID: \"ba604131-ba14-4754-8c5e-3e4fdccdfe64\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8jpzb" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.759807 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78dm8\" (UniqueName: \"kubernetes.io/projected/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-kube-api-access-78dm8\") pod \"console-f9d7485db-p8c2x\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.759832 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3ba307a-a762-4e65-aee8-d30964fadea9-config\") pod \"kube-controller-manager-operator-78b949d7b-gdpgt\" (UID: \"d3ba307a-a762-4e65-aee8-d30964fadea9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gdpgt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.759861 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/46527d26-eb38-494f-85f7-f5f56b8dfada-images\") pod \"machine-api-operator-5694c8668f-whhqx\" (UID: \"46527d26-eb38-494f-85f7-f5f56b8dfada\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-whhqx" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.759884 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b1dac22-8539-4f61-ac28-2f420c845a2a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hh662\" (UID: \"0b1dac22-8539-4f61-ac28-2f420c845a2a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hh662" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.759907 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d812eb7a-ce30-4759-a558-fff3738068b6-config\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.759932 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c16c1071-8350-4dda-b34b-183cad108e19-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ph956\" (UID: \"c16c1071-8350-4dda-b34b-183cad108e19\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ph956" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.759959 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b2a55db0-a3bd-413f-a52f-75b36654377c-trusted-ca\") pod \"ingress-operator-5b745b69d9-mn55p\" (UID: \"b2a55db0-a3bd-413f-a52f-75b36654377c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn55p" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.759985 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/145ee02a-c7aa-4990-a288-e6f69f8ab030-audit-policies\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760010 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46527d26-eb38-494f-85f7-f5f56b8dfada-config\") pod \"machine-api-operator-5694c8668f-whhqx\" (UID: \"46527d26-eb38-494f-85f7-f5f56b8dfada\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-whhqx" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760035 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d812eb7a-ce30-4759-a558-fff3738068b6-audit\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760069 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760113 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5c9fc52e-279a-44a8-a779-17966e320a93-etcd-ca\") pod \"etcd-operator-b45778765-nr4vm\" (UID: \"5c9fc52e-279a-44a8-a779-17966e320a93\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760146 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/901df0a0-1dae-41aa-8d29-0981524d9312-config\") pod \"authentication-operator-69f744f599-dwv8c\" (UID: \"901df0a0-1dae-41aa-8d29-0981524d9312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760170 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5548c900-e014-4c35-b62e-455cd82f5887-serving-cert\") pod \"controller-manager-879f6c89f-8npds\" (UID: \"5548c900-e014-4c35-b62e-455cd82f5887\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760196 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m9q5\" (UniqueName: \"kubernetes.io/projected/4c71cabc-c13f-4ae8-a704-10efee323e35-kube-api-access-4m9q5\") pod \"openshift-config-operator-7777fb866f-5fhq7\" (UID: \"4c71cabc-c13f-4ae8-a704-10efee323e35\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760222 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/bd2142a6-e9cb-432d-adf9-61b5d3f3c5b1-signing-key\") pod \"service-ca-9c57cc56f-w4pmg\" (UID: \"bd2142a6-e9cb-432d-adf9-61b5d3f3c5b1\") " pod="openshift-service-ca/service-ca-9c57cc56f-w4pmg" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760252 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5c9fc52e-279a-44a8-a779-17966e320a93-etcd-client\") pod \"etcd-operator-b45778765-nr4vm\" (UID: \"5c9fc52e-279a-44a8-a779-17966e320a93\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760439 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/46527d26-eb38-494f-85f7-f5f56b8dfada-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-whhqx\" (UID: \"46527d26-eb38-494f-85f7-f5f56b8dfada\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-whhqx" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760483 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d3ba307a-a762-4e65-aee8-d30964fadea9-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-gdpgt\" (UID: \"d3ba307a-a762-4e65-aee8-d30964fadea9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gdpgt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760514 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22f69\" (UniqueName: \"kubernetes.io/projected/901df0a0-1dae-41aa-8d29-0981524d9312-kube-api-access-22f69\") pod \"authentication-operator-69f744f599-dwv8c\" (UID: \"901df0a0-1dae-41aa-8d29-0981524d9312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760542 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c2f61301-d046-4651-b76c-9939b18fb792-audit-dir\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760567 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760592 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/af023fee-eae1-427a-8d70-487588422316-profile-collector-cert\") pod \"catalog-operator-68c6474976-5cc9p\" (UID: \"af023fee-eae1-427a-8d70-487588422316\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760621 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c71cabc-c13f-4ae8-a704-10efee323e35-serving-cert\") pod \"openshift-config-operator-7777fb866f-5fhq7\" (UID: \"4c71cabc-c13f-4ae8-a704-10efee323e35\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760645 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/145ee02a-c7aa-4990-a288-e6f69f8ab030-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760670 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b1dac22-8539-4f61-ac28-2f420c845a2a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hh662\" (UID: \"0b1dac22-8539-4f61-ac28-2f420c845a2a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hh662" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760695 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d812eb7a-ce30-4759-a558-fff3738068b6-encryption-config\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760727 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/753c599b-a401-481e-a5b3-e2dd6ed2c4e7-config\") pod \"console-operator-58897d9998-jckgt\" (UID: \"753c599b-a401-481e-a5b3-e2dd6ed2c4e7\") " pod="openshift-console-operator/console-operator-58897d9998-jckgt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760762 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760787 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d812eb7a-ce30-4759-a558-fff3738068b6-node-pullsecrets\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760820 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760848 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwk6g\" (UniqueName: \"kubernetes.io/projected/959b5ca2-d235-4457-98ec-fe077dfcdc6f-kube-api-access-wwk6g\") pod \"route-controller-manager-6576b87f9c-r5zxc\" (UID: \"959b5ca2-d235-4457-98ec-fe077dfcdc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760883 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p228b\" (UniqueName: \"kubernetes.io/projected/5c9fc52e-279a-44a8-a779-17966e320a93-kube-api-access-p228b\") pod \"etcd-operator-b45778765-nr4vm\" (UID: \"5c9fc52e-279a-44a8-a779-17966e320a93\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760910 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/145ee02a-c7aa-4990-a288-e6f69f8ab030-audit-dir\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760942 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760969 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/901df0a0-1dae-41aa-8d29-0981524d9312-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dwv8c\" (UID: \"901df0a0-1dae-41aa-8d29-0981524d9312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760976 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-trusted-ca-bundle\") pod \"console-f9d7485db-p8c2x\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761004 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/901df0a0-1dae-41aa-8d29-0981524d9312-service-ca-bundle\") pod \"authentication-operator-69f744f599-dwv8c\" (UID: \"901df0a0-1dae-41aa-8d29-0981524d9312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.760401 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-7tf92"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761296 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zchqn"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761344 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c2f61301-d046-4651-b76c-9939b18fb792-audit-dir\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761388 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d812eb7a-ce30-4759-a558-fff3738068b6-image-import-ca\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761435 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a1e4154b-2230-4499-b645-4903ea9593f1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-zcl7d\" (UID: \"a1e4154b-2230-4499-b645-4903ea9593f1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zcl7d" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761459 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/12f53d40-4cdb-41e5-b620-c85db6cb410a-metrics-tls\") pod \"dns-operator-744455d44c-z76rm\" (UID: \"12f53d40-4cdb-41e5-b620-c85db6cb410a\") " pod="openshift-dns-operator/dns-operator-744455d44c-z76rm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761482 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/145ee02a-c7aa-4990-a288-e6f69f8ab030-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761501 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwt7f\" (UniqueName: \"kubernetes.io/projected/145ee02a-c7aa-4990-a288-e6f69f8ab030-kube-api-access-qwt7f\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761522 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nm8t\" (UniqueName: \"kubernetes.io/projected/3da2af1b-ff3c-420c-af84-633eaddc9368-kube-api-access-4nm8t\") pod \"migrator-59844c95c7-6bmfr\" (UID: \"3da2af1b-ff3c-420c-af84-633eaddc9368\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bmfr" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761540 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78j88\" (UniqueName: \"kubernetes.io/projected/12f53d40-4cdb-41e5-b620-c85db6cb410a-kube-api-access-78j88\") pod \"dns-operator-744455d44c-z76rm\" (UID: \"12f53d40-4cdb-41e5-b620-c85db6cb410a\") " pod="openshift-dns-operator/dns-operator-744455d44c-z76rm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761565 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761585 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/145ee02a-c7aa-4990-a288-e6f69f8ab030-etcd-client\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761605 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761625 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5c9fc52e-279a-44a8-a779-17966e320a93-etcd-service-ca\") pod \"etcd-operator-b45778765-nr4vm\" (UID: \"5c9fc52e-279a-44a8-a779-17966e320a93\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761646 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3ba307a-a762-4e65-aee8-d30964fadea9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-gdpgt\" (UID: \"d3ba307a-a762-4e65-aee8-d30964fadea9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gdpgt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761769 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-console-serving-cert\") pod \"console-f9d7485db-p8c2x\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761794 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkkfs\" (UniqueName: \"kubernetes.io/projected/a0d34185-bf08-456c-a5fc-3f0852d35a91-kube-api-access-lkkfs\") pod \"downloads-7954f5f757-8msj4\" (UID: \"a0d34185-bf08-456c-a5fc-3f0852d35a91\") " pod="openshift-console/downloads-7954f5f757-8msj4" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761818 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkfzp\" (UniqueName: \"kubernetes.io/projected/cb832863-8ab8-459a-8d73-3c464b8caf34-kube-api-access-xkfzp\") pod \"machine-config-controller-84d6567774-wvjnl\" (UID: \"cb832863-8ab8-459a-8d73-3c464b8caf34\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wvjnl" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761839 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d812eb7a-ce30-4759-a558-fff3738068b6-etcd-serving-ca\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761858 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5548c900-e014-4c35-b62e-455cd82f5887-client-ca\") pod \"controller-manager-879f6c89f-8npds\" (UID: \"5548c900-e014-4c35-b62e-455cd82f5887\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761992 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-oauth-serving-cert\") pod \"console-f9d7485db-p8c2x\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.761880 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6342508-1473-49c4-a78f-b4e6ed9c68b0-config\") pod \"openshift-apiserver-operator-796bbdcf4f-9s2ls\" (UID: \"d6342508-1473-49c4-a78f-b4e6ed9c68b0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9s2ls" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.762516 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/145ee02a-c7aa-4990-a288-e6f69f8ab030-serving-cert\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.762541 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/753c599b-a401-481e-a5b3-e2dd6ed2c4e7-serving-cert\") pod \"console-operator-58897d9998-jckgt\" (UID: \"753c599b-a401-481e-a5b3-e2dd6ed2c4e7\") " pod="openshift-console-operator/console-operator-58897d9998-jckgt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.762563 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/959b5ca2-d235-4457-98ec-fe077dfcdc6f-client-ca\") pod \"route-controller-manager-6576b87f9c-r5zxc\" (UID: \"959b5ca2-d235-4457-98ec-fe077dfcdc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.762587 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/4c71cabc-c13f-4ae8-a704-10efee323e35-available-featuregates\") pod \"openshift-config-operator-7777fb866f-5fhq7\" (UID: \"4c71cabc-c13f-4ae8-a704-10efee323e35\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.762615 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs89d\" (UniqueName: \"kubernetes.io/projected/d812eb7a-ce30-4759-a558-fff3738068b6-kube-api-access-zs89d\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.762629 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.762640 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.762632 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.762684 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cb832863-8ab8-459a-8d73-3c464b8caf34-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wvjnl\" (UID: \"cb832863-8ab8-459a-8d73-3c464b8caf34\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wvjnl" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.762791 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.763012 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ebeee10-d90f-4672-8bfa-5f10fa929d31-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hn7lm\" (UID: \"7ebeee10-d90f-4672-8bfa-5f10fa929d31\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hn7lm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.763481 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/753c599b-a401-481e-a5b3-e2dd6ed2c4e7-config\") pod \"console-operator-58897d9998-jckgt\" (UID: \"753c599b-a401-481e-a5b3-e2dd6ed2c4e7\") " pod="openshift-console-operator/console-operator-58897d9998-jckgt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.763551 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-console-config\") pod \"console-f9d7485db-p8c2x\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.763579 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1e4154b-2230-4499-b645-4903ea9593f1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-zcl7d\" (UID: \"a1e4154b-2230-4499-b645-4903ea9593f1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zcl7d" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.763602 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c9fc52e-279a-44a8-a779-17966e320a93-config\") pod \"etcd-operator-b45778765-nr4vm\" (UID: \"5c9fc52e-279a-44a8-a779-17966e320a93\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.763630 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5548c900-e014-4c35-b62e-455cd82f5887-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8npds\" (UID: \"5548c900-e014-4c35-b62e-455cd82f5887\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.763680 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-console-oauth-config\") pod \"console-f9d7485db-p8c2x\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.763712 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.763747 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6342508-1473-49c4-a78f-b4e6ed9c68b0-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-9s2ls\" (UID: \"d6342508-1473-49c4-a78f-b4e6ed9c68b0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9s2ls" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.763776 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmk2v\" (UniqueName: \"kubernetes.io/projected/5548c900-e014-4c35-b62e-455cd82f5887-kube-api-access-fmk2v\") pod \"controller-manager-879f6c89f-8npds\" (UID: \"5548c900-e014-4c35-b62e-455cd82f5887\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.763797 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-audit-policies\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.763813 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b2a55db0-a3bd-413f-a52f-75b36654377c-metrics-tls\") pod \"ingress-operator-5b745b69d9-mn55p\" (UID: \"b2a55db0-a3bd-413f-a52f-75b36654377c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn55p" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.763843 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/901df0a0-1dae-41aa-8d29-0981524d9312-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dwv8c\" (UID: \"901df0a0-1dae-41aa-8d29-0981524d9312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.763861 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-service-ca\") pod \"console-f9d7485db-p8c2x\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.763885 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.763905 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64k2c\" (UniqueName: \"kubernetes.io/projected/b2a55db0-a3bd-413f-a52f-75b36654377c-kube-api-access-64k2c\") pod \"ingress-operator-5b745b69d9-mn55p\" (UID: \"b2a55db0-a3bd-413f-a52f-75b36654377c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn55p" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.764422 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/901df0a0-1dae-41aa-8d29-0981524d9312-service-ca-bundle\") pod \"authentication-operator-69f744f599-dwv8c\" (UID: \"901df0a0-1dae-41aa-8d29-0981524d9312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.764825 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-5n6pv"] Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.765325 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/901df0a0-1dae-41aa-8d29-0981524d9312-config\") pod \"authentication-operator-69f744f599-dwv8c\" (UID: \"901df0a0-1dae-41aa-8d29-0981524d9312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.765432 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/753c599b-a401-481e-a5b3-e2dd6ed2c4e7-trusted-ca\") pod \"console-operator-58897d9998-jckgt\" (UID: \"753c599b-a401-481e-a5b3-e2dd6ed2c4e7\") " pod="openshift-console-operator/console-operator-58897d9998-jckgt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.765460 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-5n6pv" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.765869 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/959b5ca2-d235-4457-98ec-fe077dfcdc6f-serving-cert\") pod \"route-controller-manager-6576b87f9c-r5zxc\" (UID: \"959b5ca2-d235-4457-98ec-fe077dfcdc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.765910 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/959b5ca2-d235-4457-98ec-fe077dfcdc6f-config\") pod \"route-controller-manager-6576b87f9c-r5zxc\" (UID: \"959b5ca2-d235-4457-98ec-fe077dfcdc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.765987 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.767054 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-console-config\") pod \"console-f9d7485db-p8c2x\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.767349 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.767610 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.767617 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/901df0a0-1dae-41aa-8d29-0981524d9312-serving-cert\") pod \"authentication-operator-69f744f599-dwv8c\" (UID: \"901df0a0-1dae-41aa-8d29-0981524d9312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.767688 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/af023fee-eae1-427a-8d70-487588422316-srv-cert\") pod \"catalog-operator-68c6474976-5cc9p\" (UID: \"af023fee-eae1-427a-8d70-487588422316\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.767719 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cb832863-8ab8-459a-8d73-3c464b8caf34-proxy-tls\") pod \"machine-config-controller-84d6567774-wvjnl\" (UID: \"cb832863-8ab8-459a-8d73-3c464b8caf34\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wvjnl" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.767767 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5548c900-e014-4c35-b62e-455cd82f5887-config\") pod \"controller-manager-879f6c89f-8npds\" (UID: \"5548c900-e014-4c35-b62e-455cd82f5887\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.767800 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/ba604131-ba14-4754-8c5e-3e4fdccdfe64-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8jpzb\" (UID: \"ba604131-ba14-4754-8c5e-3e4fdccdfe64\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8jpzb" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.767889 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c9fc52e-279a-44a8-a779-17966e320a93-serving-cert\") pod \"etcd-operator-b45778765-nr4vm\" (UID: \"5c9fc52e-279a-44a8-a779-17966e320a93\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.767943 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d812eb7a-ce30-4759-a558-fff3738068b6-audit-dir\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.767960 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c16c1071-8350-4dda-b34b-183cad108e19-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ph956\" (UID: \"c16c1071-8350-4dda-b34b-183cad108e19\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ph956" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.767985 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q97bs\" (UniqueName: \"kubernetes.io/projected/753c599b-a401-481e-a5b3-e2dd6ed2c4e7-kube-api-access-q97bs\") pod \"console-operator-58897d9998-jckgt\" (UID: \"753c599b-a401-481e-a5b3-e2dd6ed2c4e7\") " pod="openshift-console-operator/console-operator-58897d9998-jckgt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.768024 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c16c1071-8350-4dda-b34b-183cad108e19-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ph956\" (UID: \"c16c1071-8350-4dda-b34b-183cad108e19\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ph956" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.768228 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.768300 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.768800 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.768870 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6342508-1473-49c4-a78f-b4e6ed9c68b0-config\") pod \"openshift-apiserver-operator-796bbdcf4f-9s2ls\" (UID: \"d6342508-1473-49c4-a78f-b4e6ed9c68b0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9s2ls" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.768943 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ebeee10-d90f-4672-8bfa-5f10fa929d31-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hn7lm\" (UID: \"7ebeee10-d90f-4672-8bfa-5f10fa929d31\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hn7lm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.768983 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/959b5ca2-d235-4457-98ec-fe077dfcdc6f-client-ca\") pod \"route-controller-manager-6576b87f9c-r5zxc\" (UID: \"959b5ca2-d235-4457-98ec-fe077dfcdc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.769243 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.769311 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.769558 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-service-ca\") pod \"console-f9d7485db-p8c2x\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.769830 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-audit-policies\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.770013 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.770406 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/753c599b-a401-481e-a5b3-e2dd6ed2c4e7-serving-cert\") pod \"console-operator-58897d9998-jckgt\" (UID: \"753c599b-a401-481e-a5b3-e2dd6ed2c4e7\") " pod="openshift-console-operator/console-operator-58897d9998-jckgt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.771621 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6342508-1473-49c4-a78f-b4e6ed9c68b0-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-9s2ls\" (UID: \"d6342508-1473-49c4-a78f-b4e6ed9c68b0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9s2ls" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.771950 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.772496 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-console-serving-cert\") pod \"console-f9d7485db-p8c2x\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.772827 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-console-oauth-config\") pod \"console-f9d7485db-p8c2x\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.772974 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/ba604131-ba14-4754-8c5e-3e4fdccdfe64-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-8jpzb\" (UID: \"ba604131-ba14-4754-8c5e-3e4fdccdfe64\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8jpzb" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.785425 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.805307 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.826011 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.845270 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.865482 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.868530 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3ba307a-a762-4e65-aee8-d30964fadea9-config\") pod \"kube-controller-manager-operator-78b949d7b-gdpgt\" (UID: \"d3ba307a-a762-4e65-aee8-d30964fadea9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gdpgt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.868563 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c16c1071-8350-4dda-b34b-183cad108e19-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ph956\" (UID: \"c16c1071-8350-4dda-b34b-183cad108e19\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ph956" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.868592 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b2a55db0-a3bd-413f-a52f-75b36654377c-trusted-ca\") pod \"ingress-operator-5b745b69d9-mn55p\" (UID: \"b2a55db0-a3bd-413f-a52f-75b36654377c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn55p" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.868615 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/46527d26-eb38-494f-85f7-f5f56b8dfada-images\") pod \"machine-api-operator-5694c8668f-whhqx\" (UID: \"46527d26-eb38-494f-85f7-f5f56b8dfada\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-whhqx" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.868639 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b1dac22-8539-4f61-ac28-2f420c845a2a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hh662\" (UID: \"0b1dac22-8539-4f61-ac28-2f420c845a2a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hh662" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.868689 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d812eb7a-ce30-4759-a558-fff3738068b6-config\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.868714 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d812eb7a-ce30-4759-a558-fff3738068b6-audit\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.868761 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/145ee02a-c7aa-4990-a288-e6f69f8ab030-audit-policies\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.868783 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46527d26-eb38-494f-85f7-f5f56b8dfada-config\") pod \"machine-api-operator-5694c8668f-whhqx\" (UID: \"46527d26-eb38-494f-85f7-f5f56b8dfada\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-whhqx" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.868808 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5c9fc52e-279a-44a8-a779-17966e320a93-etcd-ca\") pod \"etcd-operator-b45778765-nr4vm\" (UID: \"5c9fc52e-279a-44a8-a779-17966e320a93\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.868832 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5548c900-e014-4c35-b62e-455cd82f5887-serving-cert\") pod \"controller-manager-879f6c89f-8npds\" (UID: \"5548c900-e014-4c35-b62e-455cd82f5887\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.868859 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m9q5\" (UniqueName: \"kubernetes.io/projected/4c71cabc-c13f-4ae8-a704-10efee323e35-kube-api-access-4m9q5\") pod \"openshift-config-operator-7777fb866f-5fhq7\" (UID: \"4c71cabc-c13f-4ae8-a704-10efee323e35\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.868883 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/bd2142a6-e9cb-432d-adf9-61b5d3f3c5b1-signing-key\") pod \"service-ca-9c57cc56f-w4pmg\" (UID: \"bd2142a6-e9cb-432d-adf9-61b5d3f3c5b1\") " pod="openshift-service-ca/service-ca-9c57cc56f-w4pmg" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.868915 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5c9fc52e-279a-44a8-a779-17966e320a93-etcd-client\") pod \"etcd-operator-b45778765-nr4vm\" (UID: \"5c9fc52e-279a-44a8-a779-17966e320a93\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.868940 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/46527d26-eb38-494f-85f7-f5f56b8dfada-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-whhqx\" (UID: \"46527d26-eb38-494f-85f7-f5f56b8dfada\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-whhqx" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.868962 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d3ba307a-a762-4e65-aee8-d30964fadea9-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-gdpgt\" (UID: \"d3ba307a-a762-4e65-aee8-d30964fadea9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gdpgt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.868985 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/af023fee-eae1-427a-8d70-487588422316-profile-collector-cert\") pod \"catalog-operator-68c6474976-5cc9p\" (UID: \"af023fee-eae1-427a-8d70-487588422316\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869009 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c71cabc-c13f-4ae8-a704-10efee323e35-serving-cert\") pod \"openshift-config-operator-7777fb866f-5fhq7\" (UID: \"4c71cabc-c13f-4ae8-a704-10efee323e35\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869034 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/145ee02a-c7aa-4990-a288-e6f69f8ab030-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869055 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b1dac22-8539-4f61-ac28-2f420c845a2a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hh662\" (UID: \"0b1dac22-8539-4f61-ac28-2f420c845a2a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hh662" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869094 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d812eb7a-ce30-4759-a558-fff3738068b6-encryption-config\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869125 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p228b\" (UniqueName: \"kubernetes.io/projected/5c9fc52e-279a-44a8-a779-17966e320a93-kube-api-access-p228b\") pod \"etcd-operator-b45778765-nr4vm\" (UID: \"5c9fc52e-279a-44a8-a779-17966e320a93\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869147 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d812eb7a-ce30-4759-a558-fff3738068b6-node-pullsecrets\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869171 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/145ee02a-c7aa-4990-a288-e6f69f8ab030-audit-dir\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869194 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d812eb7a-ce30-4759-a558-fff3738068b6-image-import-ca\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869228 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a1e4154b-2230-4499-b645-4903ea9593f1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-zcl7d\" (UID: \"a1e4154b-2230-4499-b645-4903ea9593f1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zcl7d" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869249 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/12f53d40-4cdb-41e5-b620-c85db6cb410a-metrics-tls\") pod \"dns-operator-744455d44c-z76rm\" (UID: \"12f53d40-4cdb-41e5-b620-c85db6cb410a\") " pod="openshift-dns-operator/dns-operator-744455d44c-z76rm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869274 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78j88\" (UniqueName: \"kubernetes.io/projected/12f53d40-4cdb-41e5-b620-c85db6cb410a-kube-api-access-78j88\") pod \"dns-operator-744455d44c-z76rm\" (UID: \"12f53d40-4cdb-41e5-b620-c85db6cb410a\") " pod="openshift-dns-operator/dns-operator-744455d44c-z76rm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869295 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/145ee02a-c7aa-4990-a288-e6f69f8ab030-etcd-client\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869316 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/145ee02a-c7aa-4990-a288-e6f69f8ab030-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869324 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b1dac22-8539-4f61-ac28-2f420c845a2a-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hh662\" (UID: \"0b1dac22-8539-4f61-ac28-2f420c845a2a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hh662" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869338 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwt7f\" (UniqueName: \"kubernetes.io/projected/145ee02a-c7aa-4990-a288-e6f69f8ab030-kube-api-access-qwt7f\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869361 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nm8t\" (UniqueName: \"kubernetes.io/projected/3da2af1b-ff3c-420c-af84-633eaddc9368-kube-api-access-4nm8t\") pod \"migrator-59844c95c7-6bmfr\" (UID: \"3da2af1b-ff3c-420c-af84-633eaddc9368\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bmfr" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869391 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5c9fc52e-279a-44a8-a779-17966e320a93-etcd-service-ca\") pod \"etcd-operator-b45778765-nr4vm\" (UID: \"5c9fc52e-279a-44a8-a779-17966e320a93\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869412 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3ba307a-a762-4e65-aee8-d30964fadea9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-gdpgt\" (UID: \"d3ba307a-a762-4e65-aee8-d30964fadea9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gdpgt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869435 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/145ee02a-c7aa-4990-a288-e6f69f8ab030-serving-cert\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869459 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkfzp\" (UniqueName: \"kubernetes.io/projected/cb832863-8ab8-459a-8d73-3c464b8caf34-kube-api-access-xkfzp\") pod \"machine-config-controller-84d6567774-wvjnl\" (UID: \"cb832863-8ab8-459a-8d73-3c464b8caf34\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wvjnl" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869480 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d812eb7a-ce30-4759-a558-fff3738068b6-etcd-serving-ca\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869515 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5548c900-e014-4c35-b62e-455cd82f5887-client-ca\") pod \"controller-manager-879f6c89f-8npds\" (UID: \"5548c900-e014-4c35-b62e-455cd82f5887\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869540 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/4c71cabc-c13f-4ae8-a704-10efee323e35-available-featuregates\") pod \"openshift-config-operator-7777fb866f-5fhq7\" (UID: \"4c71cabc-c13f-4ae8-a704-10efee323e35\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869566 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs89d\" (UniqueName: \"kubernetes.io/projected/d812eb7a-ce30-4759-a558-fff3738068b6-kube-api-access-zs89d\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869596 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/46527d26-eb38-494f-85f7-f5f56b8dfada-images\") pod \"machine-api-operator-5694c8668f-whhqx\" (UID: \"46527d26-eb38-494f-85f7-f5f56b8dfada\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-whhqx" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869601 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/145ee02a-c7aa-4990-a288-e6f69f8ab030-audit-policies\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869632 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5c9fc52e-279a-44a8-a779-17966e320a93-etcd-ca\") pod \"etcd-operator-b45778765-nr4vm\" (UID: \"5c9fc52e-279a-44a8-a779-17966e320a93\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869609 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cb832863-8ab8-459a-8d73-3c464b8caf34-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wvjnl\" (UID: \"cb832863-8ab8-459a-8d73-3c464b8caf34\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wvjnl" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869686 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5548c900-e014-4c35-b62e-455cd82f5887-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8npds\" (UID: \"5548c900-e014-4c35-b62e-455cd82f5887\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869884 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d812eb7a-ce30-4759-a558-fff3738068b6-node-pullsecrets\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.869918 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/145ee02a-c7aa-4990-a288-e6f69f8ab030-audit-dir\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.870528 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/4c71cabc-c13f-4ae8-a704-10efee323e35-available-featuregates\") pod \"openshift-config-operator-7777fb866f-5fhq7\" (UID: \"4c71cabc-c13f-4ae8-a704-10efee323e35\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.870548 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cb832863-8ab8-459a-8d73-3c464b8caf34-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-wvjnl\" (UID: \"cb832863-8ab8-459a-8d73-3c464b8caf34\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wvjnl" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.870725 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5548c900-e014-4c35-b62e-455cd82f5887-client-ca\") pod \"controller-manager-879f6c89f-8npds\" (UID: \"5548c900-e014-4c35-b62e-455cd82f5887\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.870750 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/145ee02a-c7aa-4990-a288-e6f69f8ab030-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.870779 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1e4154b-2230-4499-b645-4903ea9593f1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-zcl7d\" (UID: \"a1e4154b-2230-4499-b645-4903ea9593f1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zcl7d" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.870804 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c9fc52e-279a-44a8-a779-17966e320a93-config\") pod \"etcd-operator-b45778765-nr4vm\" (UID: \"5c9fc52e-279a-44a8-a779-17966e320a93\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.871444 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5c9fc52e-279a-44a8-a779-17966e320a93-etcd-service-ca\") pod \"etcd-operator-b45778765-nr4vm\" (UID: \"5c9fc52e-279a-44a8-a779-17966e320a93\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.871532 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46527d26-eb38-494f-85f7-f5f56b8dfada-config\") pod \"machine-api-operator-5694c8668f-whhqx\" (UID: \"46527d26-eb38-494f-85f7-f5f56b8dfada\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-whhqx" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.871634 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5548c900-e014-4c35-b62e-455cd82f5887-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-8npds\" (UID: \"5548c900-e014-4c35-b62e-455cd82f5887\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.871721 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmk2v\" (UniqueName: \"kubernetes.io/projected/5548c900-e014-4c35-b62e-455cd82f5887-kube-api-access-fmk2v\") pod \"controller-manager-879f6c89f-8npds\" (UID: \"5548c900-e014-4c35-b62e-455cd82f5887\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.871823 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b2a55db0-a3bd-413f-a52f-75b36654377c-metrics-tls\") pod \"ingress-operator-5b745b69d9-mn55p\" (UID: \"b2a55db0-a3bd-413f-a52f-75b36654377c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn55p" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.871883 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64k2c\" (UniqueName: \"kubernetes.io/projected/b2a55db0-a3bd-413f-a52f-75b36654377c-kube-api-access-64k2c\") pod \"ingress-operator-5b745b69d9-mn55p\" (UID: \"b2a55db0-a3bd-413f-a52f-75b36654377c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn55p" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.871911 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/af023fee-eae1-427a-8d70-487588422316-srv-cert\") pod \"catalog-operator-68c6474976-5cc9p\" (UID: \"af023fee-eae1-427a-8d70-487588422316\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.871933 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cb832863-8ab8-459a-8d73-3c464b8caf34-proxy-tls\") pod \"machine-config-controller-84d6567774-wvjnl\" (UID: \"cb832863-8ab8-459a-8d73-3c464b8caf34\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wvjnl" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.871955 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5548c900-e014-4c35-b62e-455cd82f5887-config\") pod \"controller-manager-879f6c89f-8npds\" (UID: \"5548c900-e014-4c35-b62e-455cd82f5887\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.871978 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c16c1071-8350-4dda-b34b-183cad108e19-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ph956\" (UID: \"c16c1071-8350-4dda-b34b-183cad108e19\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ph956" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.872012 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c9fc52e-279a-44a8-a779-17966e320a93-serving-cert\") pod \"etcd-operator-b45778765-nr4vm\" (UID: \"5c9fc52e-279a-44a8-a779-17966e320a93\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.872033 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d812eb7a-ce30-4759-a558-fff3738068b6-audit-dir\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.872053 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c16c1071-8350-4dda-b34b-183cad108e19-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ph956\" (UID: \"c16c1071-8350-4dda-b34b-183cad108e19\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ph956" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.872075 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/145ee02a-c7aa-4990-a288-e6f69f8ab030-encryption-config\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.872122 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2ktb\" (UniqueName: \"kubernetes.io/projected/bd2142a6-e9cb-432d-adf9-61b5d3f3c5b1-kube-api-access-m2ktb\") pod \"service-ca-9c57cc56f-w4pmg\" (UID: \"bd2142a6-e9cb-432d-adf9-61b5d3f3c5b1\") " pod="openshift-service-ca/service-ca-9c57cc56f-w4pmg" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.872124 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c9fc52e-279a-44a8-a779-17966e320a93-config\") pod \"etcd-operator-b45778765-nr4vm\" (UID: \"5c9fc52e-279a-44a8-a779-17966e320a93\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.872391 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/bd2142a6-e9cb-432d-adf9-61b5d3f3c5b1-signing-cabundle\") pod \"service-ca-9c57cc56f-w4pmg\" (UID: \"bd2142a6-e9cb-432d-adf9-61b5d3f3c5b1\") " pod="openshift-service-ca/service-ca-9c57cc56f-w4pmg" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.872444 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxj8b\" (UniqueName: \"kubernetes.io/projected/46527d26-eb38-494f-85f7-f5f56b8dfada-kube-api-access-qxj8b\") pod \"machine-api-operator-5694c8668f-whhqx\" (UID: \"46527d26-eb38-494f-85f7-f5f56b8dfada\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-whhqx" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.872478 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b2a55db0-a3bd-413f-a52f-75b36654377c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mn55p\" (UID: \"b2a55db0-a3bd-413f-a52f-75b36654377c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn55p" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.872606 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24q7q\" (UniqueName: \"kubernetes.io/projected/af023fee-eae1-427a-8d70-487588422316-kube-api-access-24q7q\") pod \"catalog-operator-68c6474976-5cc9p\" (UID: \"af023fee-eae1-427a-8d70-487588422316\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.872620 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/145ee02a-c7aa-4990-a288-e6f69f8ab030-etcd-client\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.872630 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzlhf\" (UniqueName: \"kubernetes.io/projected/0b1dac22-8539-4f61-ac28-2f420c845a2a-kube-api-access-hzlhf\") pod \"kube-storage-version-migrator-operator-b67b599dd-hh662\" (UID: \"0b1dac22-8539-4f61-ac28-2f420c845a2a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hh662" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.872686 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d812eb7a-ce30-4759-a558-fff3738068b6-etcd-client\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.872709 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d812eb7a-ce30-4759-a558-fff3738068b6-serving-cert\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.872741 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d812eb7a-ce30-4759-a558-fff3738068b6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.872774 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1e4154b-2230-4499-b645-4903ea9593f1-config\") pod \"kube-apiserver-operator-766d6c64bb-zcl7d\" (UID: \"a1e4154b-2230-4499-b645-4903ea9593f1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zcl7d" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.873032 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/145ee02a-c7aa-4990-a288-e6f69f8ab030-serving-cert\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.873164 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d812eb7a-ce30-4759-a558-fff3738068b6-audit-dir\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.873520 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c16c1071-8350-4dda-b34b-183cad108e19-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ph956\" (UID: \"c16c1071-8350-4dda-b34b-183cad108e19\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ph956" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.873629 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1e4154b-2230-4499-b645-4903ea9593f1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-zcl7d\" (UID: \"a1e4154b-2230-4499-b645-4903ea9593f1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zcl7d" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.873692 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b1dac22-8539-4f61-ac28-2f420c845a2a-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hh662\" (UID: \"0b1dac22-8539-4f61-ac28-2f420c845a2a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hh662" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.873780 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1e4154b-2230-4499-b645-4903ea9593f1-config\") pod \"kube-apiserver-operator-766d6c64bb-zcl7d\" (UID: \"a1e4154b-2230-4499-b645-4903ea9593f1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zcl7d" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.874244 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5c9fc52e-279a-44a8-a779-17966e320a93-etcd-client\") pod \"etcd-operator-b45778765-nr4vm\" (UID: \"5c9fc52e-279a-44a8-a779-17966e320a93\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.874548 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5548c900-e014-4c35-b62e-455cd82f5887-config\") pod \"controller-manager-879f6c89f-8npds\" (UID: \"5548c900-e014-4c35-b62e-455cd82f5887\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.874922 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/12f53d40-4cdb-41e5-b620-c85db6cb410a-metrics-tls\") pod \"dns-operator-744455d44c-z76rm\" (UID: \"12f53d40-4cdb-41e5-b620-c85db6cb410a\") " pod="openshift-dns-operator/dns-operator-744455d44c-z76rm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.875123 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c16c1071-8350-4dda-b34b-183cad108e19-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ph956\" (UID: \"c16c1071-8350-4dda-b34b-183cad108e19\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ph956" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.875382 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5548c900-e014-4c35-b62e-455cd82f5887-serving-cert\") pod \"controller-manager-879f6c89f-8npds\" (UID: \"5548c900-e014-4c35-b62e-455cd82f5887\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.875595 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c9fc52e-279a-44a8-a779-17966e320a93-serving-cert\") pod \"etcd-operator-b45778765-nr4vm\" (UID: \"5c9fc52e-279a-44a8-a779-17966e320a93\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.875754 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/46527d26-eb38-494f-85f7-f5f56b8dfada-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-whhqx\" (UID: \"46527d26-eb38-494f-85f7-f5f56b8dfada\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-whhqx" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.877165 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/145ee02a-c7aa-4990-a288-e6f69f8ab030-encryption-config\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.884960 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.906300 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.911691 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/145ee02a-c7aa-4990-a288-e6f69f8ab030-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.925484 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.929434 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3ba307a-a762-4e65-aee8-d30964fadea9-config\") pod \"kube-controller-manager-operator-78b949d7b-gdpgt\" (UID: \"d3ba307a-a762-4e65-aee8-d30964fadea9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gdpgt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.945198 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.954048 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3ba307a-a762-4e65-aee8-d30964fadea9-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-gdpgt\" (UID: \"d3ba307a-a762-4e65-aee8-d30964fadea9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gdpgt" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.964795 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 27 07:44:44 crc kubenswrapper[4934]: I1227 07:44:44.985870 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.006258 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.025856 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.045155 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.055498 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c71cabc-c13f-4ae8-a704-10efee323e35-serving-cert\") pod \"openshift-config-operator-7777fb866f-5fhq7\" (UID: \"4c71cabc-c13f-4ae8-a704-10efee323e35\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.066113 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.086321 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.106522 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.126706 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.146356 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.158020 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cb832863-8ab8-459a-8d73-3c464b8caf34-proxy-tls\") pod \"machine-config-controller-84d6567774-wvjnl\" (UID: \"cb832863-8ab8-459a-8d73-3c464b8caf34\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wvjnl" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.165904 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.185473 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.199345 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d812eb7a-ce30-4759-a558-fff3738068b6-etcd-client\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.206431 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.217828 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d812eb7a-ce30-4759-a558-fff3738068b6-serving-cert\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.226036 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.234535 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d812eb7a-ce30-4759-a558-fff3738068b6-encryption-config\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.246337 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.249854 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d812eb7a-ce30-4759-a558-fff3738068b6-config\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.265723 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.271715 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d812eb7a-ce30-4759-a558-fff3738068b6-audit\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.279190 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:45 crc kubenswrapper[4934]: E1227 07:44:45.279461 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:46:47.279378673 +0000 UTC m=+268.099819317 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.279584 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.279657 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.280783 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.284720 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.285706 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.291661 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d812eb7a-ce30-4759-a558-fff3738068b6-etcd-serving-ca\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.306709 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.312252 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d812eb7a-ce30-4759-a558-fff3738068b6-image-import-ca\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.330154 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.330223 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.340079 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.345975 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d812eb7a-ce30-4759-a558-fff3738068b6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.346066 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.365999 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.380939 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.381352 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.386418 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.387471 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.387580 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.398450 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b2a55db0-a3bd-413f-a52f-75b36654377c-metrics-tls\") pod \"ingress-operator-5b745b69d9-mn55p\" (UID: \"b2a55db0-a3bd-413f-a52f-75b36654377c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn55p" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.400999 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.414806 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.420471 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.425897 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.431653 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b2a55db0-a3bd-413f-a52f-75b36654377c-trusted-ca\") pod \"ingress-operator-5b745b69d9-mn55p\" (UID: \"b2a55db0-a3bd-413f-a52f-75b36654377c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn55p" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.437770 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.446831 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.467291 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.486783 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.495190 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/bd2142a6-e9cb-432d-adf9-61b5d3f3c5b1-signing-cabundle\") pod \"service-ca-9c57cc56f-w4pmg\" (UID: \"bd2142a6-e9cb-432d-adf9-61b5d3f3c5b1\") " pod="openshift-service-ca/service-ca-9c57cc56f-w4pmg" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.527292 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.540338 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/bd2142a6-e9cb-432d-adf9-61b5d3f3c5b1-signing-key\") pod \"service-ca-9c57cc56f-w4pmg\" (UID: \"bd2142a6-e9cb-432d-adf9-61b5d3f3c5b1\") " pod="openshift-service-ca/service-ca-9c57cc56f-w4pmg" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.547816 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.565518 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.586318 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.607201 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.626245 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.637224 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/af023fee-eae1-427a-8d70-487588422316-srv-cert\") pod \"catalog-operator-68c6474976-5cc9p\" (UID: \"af023fee-eae1-427a-8d70-487588422316\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.647256 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.663460 4934 request.go:700] Waited for 1.014888862s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dpprof-cert&limit=500&resourceVersion=0 Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.664701 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.674677 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/af023fee-eae1-427a-8d70-487588422316-profile-collector-cert\") pod \"catalog-operator-68c6474976-5cc9p\" (UID: \"af023fee-eae1-427a-8d70-487588422316\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.686232 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 27 07:44:45 crc kubenswrapper[4934]: W1227 07:44:45.696831 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-7510d7f0862c694385cab278fa9248d491cf8c3d1a7cb0fe3238d1ecc27cc192 WatchSource:0}: Error finding container 7510d7f0862c694385cab278fa9248d491cf8c3d1a7cb0fe3238d1ecc27cc192: Status 404 returned error can't find the container with id 7510d7f0862c694385cab278fa9248d491cf8c3d1a7cb0fe3238d1ecc27cc192 Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.705218 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.725891 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.745995 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.766933 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.786796 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.807465 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.836187 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.845295 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.866243 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.905680 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.926480 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.945076 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.986314 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 27 07:44:45 crc kubenswrapper[4934]: I1227 07:44:45.990200 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.014264 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.025891 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.045517 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.065996 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.085454 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.105741 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.125811 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.150489 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.166662 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.186054 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.205226 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.226030 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.246243 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.262772 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"9e52636a542def448dd6b22574423bda7c225cbc754a251cf17f370f9bfebb28"} Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.262868 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a72b5c3ef844e167ca720b8c93bcd3235c7619b5531cf48aef6a8d5d9e8bb145"} Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.264969 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"e7adb2362470c7c78b8faf8bad3e08268e07b1984aa970b1a339902757a6a8f2"} Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.265035 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"74672663a01260abd42c35fe2ce02179ef706aea0d59e33e8800bc5a0223d1fc"} Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.265342 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.265777 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.267389 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"073ce01e3ed89bc44291b0b6162257c3cd6de19b02893bce76297ab4e407c651"} Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.267459 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7510d7f0862c694385cab278fa9248d491cf8c3d1a7cb0fe3238d1ecc27cc192"} Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.286990 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.306068 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.325803 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.345650 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.365580 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.386012 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.405693 4934 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.426057 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.446267 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.465607 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.486320 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.505631 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.556256 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpk5n\" (UniqueName: \"kubernetes.io/projected/c2f61301-d046-4651-b76c-9939b18fb792-kube-api-access-jpk5n\") pod \"oauth-openshift-558db77b4-j5cpz\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.564141 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn7p8\" (UniqueName: \"kubernetes.io/projected/7ebeee10-d90f-4672-8bfa-5f10fa929d31-kube-api-access-cn7p8\") pod \"openshift-controller-manager-operator-756b6f6bc6-hn7lm\" (UID: \"7ebeee10-d90f-4672-8bfa-5f10fa929d31\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hn7lm" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.586003 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22f69\" (UniqueName: \"kubernetes.io/projected/901df0a0-1dae-41aa-8d29-0981524d9312-kube-api-access-22f69\") pod \"authentication-operator-69f744f599-dwv8c\" (UID: \"901df0a0-1dae-41aa-8d29-0981524d9312\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.603222 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcpfs\" (UniqueName: \"kubernetes.io/projected/ba604131-ba14-4754-8c5e-3e4fdccdfe64-kube-api-access-zcpfs\") pod \"cluster-samples-operator-665b6dd947-8jpzb\" (UID: \"ba604131-ba14-4754-8c5e-3e4fdccdfe64\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8jpzb" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.649634 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwk6g\" (UniqueName: \"kubernetes.io/projected/959b5ca2-d235-4457-98ec-fe077dfcdc6f-kube-api-access-wwk6g\") pod \"route-controller-manager-6576b87f9c-r5zxc\" (UID: \"959b5ca2-d235-4457-98ec-fe077dfcdc6f\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.654190 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78dm8\" (UniqueName: \"kubernetes.io/projected/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-kube-api-access-78dm8\") pod \"console-f9d7485db-p8c2x\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.664575 4934 request.go:700] Waited for 1.898881753s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-server-dockercfg-qx5rd&limit=500&resourceVersion=0 Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.666426 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.670428 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.673401 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwq7v\" (UniqueName: \"kubernetes.io/projected/d6342508-1473-49c4-a78f-b4e6ed9c68b0-kube-api-access-mwq7v\") pod \"openshift-apiserver-operator-796bbdcf4f-9s2ls\" (UID: \"d6342508-1473-49c4-a78f-b4e6ed9c68b0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9s2ls" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.687537 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.706408 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.734940 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8jpzb" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.744686 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.750256 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkkfs\" (UniqueName: \"kubernetes.io/projected/a0d34185-bf08-456c-a5fc-3f0852d35a91-kube-api-access-lkkfs\") pod \"downloads-7954f5f757-8msj4\" (UID: \"a0d34185-bf08-456c-a5fc-3f0852d35a91\") " pod="openshift-console/downloads-7954f5f757-8msj4" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.757000 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9s2ls" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.780671 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.785025 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q97bs\" (UniqueName: \"kubernetes.io/projected/753c599b-a401-481e-a5b3-e2dd6ed2c4e7-kube-api-access-q97bs\") pod \"console-operator-58897d9998-jckgt\" (UID: \"753c599b-a401-481e-a5b3-e2dd6ed2c4e7\") " pod="openshift-console-operator/console-operator-58897d9998-jckgt" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.788828 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c16c1071-8350-4dda-b34b-183cad108e19-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ph956\" (UID: \"c16c1071-8350-4dda-b34b-183cad108e19\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ph956" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.790803 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8msj4" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.799489 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.806452 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hn7lm" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.815406 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p228b\" (UniqueName: \"kubernetes.io/projected/5c9fc52e-279a-44a8-a779-17966e320a93-kube-api-access-p228b\") pod \"etcd-operator-b45778765-nr4vm\" (UID: \"5c9fc52e-279a-44a8-a779-17966e320a93\") " pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.829598 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nm8t\" (UniqueName: \"kubernetes.io/projected/3da2af1b-ff3c-420c-af84-633eaddc9368-kube-api-access-4nm8t\") pod \"migrator-59844c95c7-6bmfr\" (UID: \"3da2af1b-ff3c-420c-af84-633eaddc9368\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bmfr" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.832324 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.844523 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwt7f\" (UniqueName: \"kubernetes.io/projected/145ee02a-c7aa-4990-a288-e6f69f8ab030-kube-api-access-qwt7f\") pod \"apiserver-7bbb656c7d-hzmws\" (UID: \"145ee02a-c7aa-4990-a288-e6f69f8ab030\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.868435 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a1e4154b-2230-4499-b645-4903ea9593f1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-zcl7d\" (UID: \"a1e4154b-2230-4499-b645-4903ea9593f1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zcl7d" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.887215 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkfzp\" (UniqueName: \"kubernetes.io/projected/cb832863-8ab8-459a-8d73-3c464b8caf34-kube-api-access-xkfzp\") pod \"machine-config-controller-84d6567774-wvjnl\" (UID: \"cb832863-8ab8-459a-8d73-3c464b8caf34\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wvjnl" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.903252 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78j88\" (UniqueName: \"kubernetes.io/projected/12f53d40-4cdb-41e5-b620-c85db6cb410a-kube-api-access-78j88\") pod \"dns-operator-744455d44c-z76rm\" (UID: \"12f53d40-4cdb-41e5-b620-c85db6cb410a\") " pod="openshift-dns-operator/dns-operator-744455d44c-z76rm" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.909946 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ph956" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.922580 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.928681 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs89d\" (UniqueName: \"kubernetes.io/projected/d812eb7a-ce30-4759-a558-fff3738068b6-kube-api-access-zs89d\") pod \"apiserver-76f77b778f-8s8bp\" (UID: \"d812eb7a-ce30-4759-a558-fff3738068b6\") " pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.935188 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dwv8c"] Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.948855 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d3ba307a-a762-4e65-aee8-d30964fadea9-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-gdpgt\" (UID: \"d3ba307a-a762-4e65-aee8-d30964fadea9\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gdpgt" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.952821 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.958146 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wvjnl" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.966880 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m9q5\" (UniqueName: \"kubernetes.io/projected/4c71cabc-c13f-4ae8-a704-10efee323e35-kube-api-access-4m9q5\") pod \"openshift-config-operator-7777fb866f-5fhq7\" (UID: \"4c71cabc-c13f-4ae8-a704-10efee323e35\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" Dec 27 07:44:46 crc kubenswrapper[4934]: I1227 07:44:46.979787 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmk2v\" (UniqueName: \"kubernetes.io/projected/5548c900-e014-4c35-b62e-455cd82f5887-kube-api-access-fmk2v\") pod \"controller-manager-879f6c89f-8npds\" (UID: \"5548c900-e014-4c35-b62e-455cd82f5887\") " pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.005123 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64k2c\" (UniqueName: \"kubernetes.io/projected/b2a55db0-a3bd-413f-a52f-75b36654377c-kube-api-access-64k2c\") pod \"ingress-operator-5b745b69d9-mn55p\" (UID: \"b2a55db0-a3bd-413f-a52f-75b36654377c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn55p" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.005403 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-jckgt" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.014370 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bmfr" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.018458 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-p8c2x"] Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.034773 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzlhf\" (UniqueName: \"kubernetes.io/projected/0b1dac22-8539-4f61-ac28-2f420c845a2a-kube-api-access-hzlhf\") pod \"kube-storage-version-migrator-operator-b67b599dd-hh662\" (UID: \"0b1dac22-8539-4f61-ac28-2f420c845a2a\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hh662" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.040534 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxj8b\" (UniqueName: \"kubernetes.io/projected/46527d26-eb38-494f-85f7-f5f56b8dfada-kube-api-access-qxj8b\") pod \"machine-api-operator-5694c8668f-whhqx\" (UID: \"46527d26-eb38-494f-85f7-f5f56b8dfada\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-whhqx" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.062363 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b2a55db0-a3bd-413f-a52f-75b36654377c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-mn55p\" (UID: \"b2a55db0-a3bd-413f-a52f-75b36654377c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn55p" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.087813 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24q7q\" (UniqueName: \"kubernetes.io/projected/af023fee-eae1-427a-8d70-487588422316-kube-api-access-24q7q\") pod \"catalog-operator-68c6474976-5cc9p\" (UID: \"af023fee-eae1-427a-8d70-487588422316\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.107792 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2ktb\" (UniqueName: \"kubernetes.io/projected/bd2142a6-e9cb-432d-adf9-61b5d3f3c5b1-kube-api-access-m2ktb\") pod \"service-ca-9c57cc56f-w4pmg\" (UID: \"bd2142a6-e9cb-432d-adf9-61b5d3f3c5b1\") " pod="openshift-service-ca/service-ca-9c57cc56f-w4pmg" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.118293 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-whhqx" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.139181 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zcl7d" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.145573 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hh662" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.157584 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9s2ls"] Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.197995 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-z76rm" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.211053 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b74aaf8a-71e4-4b9a-8212-c0f1cb2fd01b-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ndhjw\" (UID: \"b74aaf8a-71e4-4b9a-8212-c0f1cb2fd01b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ndhjw" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.211131 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kdpx\" (UniqueName: \"kubernetes.io/projected/9ef59357-6b1b-491e-a43d-3d436b09d4ff-kube-api-access-5kdpx\") pod \"cluster-image-registry-operator-dc59b4c8b-22jvt\" (UID: \"9ef59357-6b1b-491e-a43d-3d436b09d4ff\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22jvt" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.211153 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/df922398-eb58-4f23-bfab-cfe9b90908c6-registry-certificates\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.211178 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/df922398-eb58-4f23-bfab-cfe9b90908c6-bound-sa-token\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.211194 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/585041c4-cd72-4116-b7cf-5801881f349c-auth-proxy-config\") pod \"machine-approver-56656f9798-2zfd9\" (UID: \"585041c4-cd72-4116-b7cf-5801881f349c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2zfd9" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.211215 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9q69n\" (UniqueName: \"kubernetes.io/projected/df922398-eb58-4f23-bfab-cfe9b90908c6-kube-api-access-9q69n\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.211232 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/585041c4-cd72-4116-b7cf-5801881f349c-config\") pod \"machine-approver-56656f9798-2zfd9\" (UID: \"585041c4-cd72-4116-b7cf-5801881f349c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2zfd9" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.211251 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9ef59357-6b1b-491e-a43d-3d436b09d4ff-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-22jvt\" (UID: \"9ef59357-6b1b-491e-a43d-3d436b09d4ff\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22jvt" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.211282 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.211304 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/585041c4-cd72-4116-b7cf-5801881f349c-machine-approver-tls\") pod \"machine-approver-56656f9798-2zfd9\" (UID: \"585041c4-cd72-4116-b7cf-5801881f349c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2zfd9" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.211321 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/df922398-eb58-4f23-bfab-cfe9b90908c6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.211347 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9ef59357-6b1b-491e-a43d-3d436b09d4ff-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-22jvt\" (UID: \"9ef59357-6b1b-491e-a43d-3d436b09d4ff\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22jvt" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.211376 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/df922398-eb58-4f23-bfab-cfe9b90908c6-trusted-ca\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.211401 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j25rp\" (UniqueName: \"kubernetes.io/projected/b74aaf8a-71e4-4b9a-8212-c0f1cb2fd01b-kube-api-access-j25rp\") pod \"multus-admission-controller-857f4d67dd-ndhjw\" (UID: \"b74aaf8a-71e4-4b9a-8212-c0f1cb2fd01b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ndhjw" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.211420 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sklws\" (UniqueName: \"kubernetes.io/projected/585041c4-cd72-4116-b7cf-5801881f349c-kube-api-access-sklws\") pod \"machine-approver-56656f9798-2zfd9\" (UID: \"585041c4-cd72-4116-b7cf-5801881f349c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2zfd9" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.211438 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9ef59357-6b1b-491e-a43d-3d436b09d4ff-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-22jvt\" (UID: \"9ef59357-6b1b-491e-a43d-3d436b09d4ff\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22jvt" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.211456 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/df922398-eb58-4f23-bfab-cfe9b90908c6-registry-tls\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.211471 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/df922398-eb58-4f23-bfab-cfe9b90908c6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: E1227 07:44:47.211794 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:47.711780964 +0000 UTC m=+148.532221558 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.215363 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.230327 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gdpgt" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.238314 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.267494 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn55p" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.273640 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" event={"ID":"901df0a0-1dae-41aa-8d29-0981524d9312","Type":"ContainerStarted","Data":"6f0be99ac7d05c0deb4dec5be7ba55a7d53e51061240102f1650ebe6d8241d1a"} Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.275486 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9s2ls" event={"ID":"d6342508-1473-49c4-a78f-b4e6ed9c68b0","Type":"ContainerStarted","Data":"39b8e564a2eb9e0efe7f703ce0355b5dc0f92fc73d33a784e2c1757368d7cb4f"} Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.276702 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-p8c2x" event={"ID":"091c02c2-ae22-44e8-bf6b-64b69d6d9d83","Type":"ContainerStarted","Data":"947b1aad9814ff50655460b476d68e661d3b2a458cec61bdef9671bd855af5dd"} Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.277014 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-w4pmg" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.297097 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc"] Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.302746 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8jpzb"] Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.304923 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.312536 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.312702 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/873d8371-2f45-409a-b18d-f4c7dbdabeea-apiservice-cert\") pod \"packageserver-d55dfcdfc-t6q48\" (UID: \"873d8371-2f45-409a-b18d-f4c7dbdabeea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.312746 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b74aaf8a-71e4-4b9a-8212-c0f1cb2fd01b-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ndhjw\" (UID: \"b74aaf8a-71e4-4b9a-8212-c0f1cb2fd01b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ndhjw" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.312767 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfmw2\" (UniqueName: \"kubernetes.io/projected/9552ddff-a57b-4311-a6d5-552a8b8efe19-kube-api-access-xfmw2\") pod \"ingress-canary-7tf92\" (UID: \"9552ddff-a57b-4311-a6d5-552a8b8efe19\") " pod="openshift-ingress-canary/ingress-canary-7tf92" Dec 27 07:44:47 crc kubenswrapper[4934]: E1227 07:44:47.312818 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:47.812803584 +0000 UTC m=+148.633244178 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.312863 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4ea539b6-ffe0-4393-af8d-62579266697c-srv-cert\") pod \"olm-operator-6b444d44fb-r6jld\" (UID: \"4ea539b6-ffe0-4393-af8d-62579266697c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.312931 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ac1de940-6952-466a-be96-52c43b1a90a2-default-certificate\") pod \"router-default-5444994796-cpts5\" (UID: \"ac1de940-6952-466a-be96-52c43b1a90a2\") " pod="openshift-ingress/router-default-5444994796-cpts5" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.312947 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a0c610b9-d890-49ee-b321-a3f1efba4b0a-plugins-dir\") pod \"csi-hostpathplugin-5s9zf\" (UID: \"a0c610b9-d890-49ee-b321-a3f1efba4b0a\") " pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.312961 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jctwn\" (UniqueName: \"kubernetes.io/projected/4ea539b6-ffe0-4393-af8d-62579266697c-kube-api-access-jctwn\") pod \"olm-operator-6b444d44fb-r6jld\" (UID: \"4ea539b6-ffe0-4393-af8d-62579266697c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.312996 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kdpx\" (UniqueName: \"kubernetes.io/projected/9ef59357-6b1b-491e-a43d-3d436b09d4ff-kube-api-access-5kdpx\") pod \"cluster-image-registry-operator-dc59b4c8b-22jvt\" (UID: \"9ef59357-6b1b-491e-a43d-3d436b09d4ff\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22jvt" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313013 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a0c610b9-d890-49ee-b321-a3f1efba4b0a-registration-dir\") pod \"csi-hostpathplugin-5s9zf\" (UID: \"a0c610b9-d890-49ee-b321-a3f1efba4b0a\") " pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313040 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/df922398-eb58-4f23-bfab-cfe9b90908c6-registry-certificates\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313056 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnzjz\" (UniqueName: \"kubernetes.io/projected/873d8371-2f45-409a-b18d-f4c7dbdabeea-kube-api-access-gnzjz\") pod \"packageserver-d55dfcdfc-t6q48\" (UID: \"873d8371-2f45-409a-b18d-f4c7dbdabeea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313115 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/df922398-eb58-4f23-bfab-cfe9b90908c6-bound-sa-token\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313147 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce5a8998-bb68-4764-aa7d-a47ec23c4514-config-volume\") pod \"collect-profiles-29447010-fl8rj\" (UID: \"ce5a8998-bb68-4764-aa7d-a47ec23c4514\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313164 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/585041c4-cd72-4116-b7cf-5801881f349c-auth-proxy-config\") pod \"machine-approver-56656f9798-2zfd9\" (UID: \"585041c4-cd72-4116-b7cf-5801881f349c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2zfd9" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313181 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc6ns\" (UniqueName: \"kubernetes.io/projected/ac1de940-6952-466a-be96-52c43b1a90a2-kube-api-access-mc6ns\") pod \"router-default-5444994796-cpts5\" (UID: \"ac1de940-6952-466a-be96-52c43b1a90a2\") " pod="openshift-ingress/router-default-5444994796-cpts5" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313199 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce5a8998-bb68-4764-aa7d-a47ec23c4514-secret-volume\") pod \"collect-profiles-29447010-fl8rj\" (UID: \"ce5a8998-bb68-4764-aa7d-a47ec23c4514\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313215 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9552ddff-a57b-4311-a6d5-552a8b8efe19-cert\") pod \"ingress-canary-7tf92\" (UID: \"9552ddff-a57b-4311-a6d5-552a8b8efe19\") " pod="openshift-ingress-canary/ingress-canary-7tf92" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313337 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9q69n\" (UniqueName: \"kubernetes.io/projected/df922398-eb58-4f23-bfab-cfe9b90908c6-kube-api-access-9q69n\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313375 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2eec2888-f042-4de8-a6f7-b2bb2bb3db14-serving-cert\") pod \"service-ca-operator-777779d784-spr46\" (UID: \"2eec2888-f042-4de8-a6f7-b2bb2bb3db14\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-spr46" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313391 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/585041c4-cd72-4116-b7cf-5801881f349c-config\") pod \"machine-approver-56656f9798-2zfd9\" (UID: \"585041c4-cd72-4116-b7cf-5801881f349c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2zfd9" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313414 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ac1de940-6952-466a-be96-52c43b1a90a2-metrics-certs\") pod \"router-default-5444994796-cpts5\" (UID: \"ac1de940-6952-466a-be96-52c43b1a90a2\") " pod="openshift-ingress/router-default-5444994796-cpts5" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313452 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9ef59357-6b1b-491e-a43d-3d436b09d4ff-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-22jvt\" (UID: \"9ef59357-6b1b-491e-a43d-3d436b09d4ff\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22jvt" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313540 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftz4h\" (UniqueName: \"kubernetes.io/projected/941a07b0-8984-43f4-8fcd-6cdc8c88f14a-kube-api-access-ftz4h\") pod \"marketplace-operator-79b997595-xxrp6\" (UID: \"941a07b0-8984-43f4-8fcd-6cdc8c88f14a\") " pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313584 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a0c610b9-d890-49ee-b321-a3f1efba4b0a-mountpoint-dir\") pod \"csi-hostpathplugin-5s9zf\" (UID: \"a0c610b9-d890-49ee-b321-a3f1efba4b0a\") " pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313606 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a0c610b9-d890-49ee-b321-a3f1efba4b0a-socket-dir\") pod \"csi-hostpathplugin-5s9zf\" (UID: \"a0c610b9-d890-49ee-b321-a3f1efba4b0a\") " pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313677 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac1de940-6952-466a-be96-52c43b1a90a2-service-ca-bundle\") pod \"router-default-5444994796-cpts5\" (UID: \"ac1de940-6952-466a-be96-52c43b1a90a2\") " pod="openshift-ingress/router-default-5444994796-cpts5" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313765 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26nvp\" (UniqueName: \"kubernetes.io/projected/2eec2888-f042-4de8-a6f7-b2bb2bb3db14-kube-api-access-26nvp\") pod \"service-ca-operator-777779d784-spr46\" (UID: \"2eec2888-f042-4de8-a6f7-b2bb2bb3db14\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-spr46" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313832 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/85b3562e-4570-4b9b-baa5-a20fe55f13ca-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-c2tqv\" (UID: \"85b3562e-4570-4b9b-baa5-a20fe55f13ca\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-c2tqv" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313850 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ac1de940-6952-466a-be96-52c43b1a90a2-stats-auth\") pod \"router-default-5444994796-cpts5\" (UID: \"ac1de940-6952-466a-be96-52c43b1a90a2\") " pod="openshift-ingress/router-default-5444994796-cpts5" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313875 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313891 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cee9f5f3-5282-4ec5-94a1-fd193ea66f2e-auth-proxy-config\") pod \"machine-config-operator-74547568cd-hrm6w\" (UID: \"cee9f5f3-5282-4ec5-94a1-fd193ea66f2e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hrm6w" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313931 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/585041c4-cd72-4116-b7cf-5801881f349c-machine-approver-tls\") pod \"machine-approver-56656f9798-2zfd9\" (UID: \"585041c4-cd72-4116-b7cf-5801881f349c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2zfd9" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313949 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/e06a3c18-5c45-4153-8b28-4cf918a55970-certs\") pod \"machine-config-server-5n6pv\" (UID: \"e06a3c18-5c45-4153-8b28-4cf918a55970\") " pod="openshift-machine-config-operator/machine-config-server-5n6pv" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.313989 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrs9l\" (UniqueName: \"kubernetes.io/projected/a0c610b9-d890-49ee-b321-a3f1efba4b0a-kube-api-access-lrs9l\") pod \"csi-hostpathplugin-5s9zf\" (UID: \"a0c610b9-d890-49ee-b321-a3f1efba4b0a\") " pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314004 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cee9f5f3-5282-4ec5-94a1-fd193ea66f2e-images\") pod \"machine-config-operator-74547568cd-hrm6w\" (UID: \"cee9f5f3-5282-4ec5-94a1-fd193ea66f2e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hrm6w" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314023 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/df922398-eb58-4f23-bfab-cfe9b90908c6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314039 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z78ms\" (UniqueName: \"kubernetes.io/projected/e4f56101-c810-4e1f-96e8-c7cbcb2ec719-kube-api-access-z78ms\") pod \"control-plane-machine-set-operator-78cbb6b69f-zchqn\" (UID: \"e4f56101-c810-4e1f-96e8-c7cbcb2ec719\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zchqn" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314113 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9ef59357-6b1b-491e-a43d-3d436b09d4ff-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-22jvt\" (UID: \"9ef59357-6b1b-491e-a43d-3d436b09d4ff\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22jvt" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314161 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/941a07b0-8984-43f4-8fcd-6cdc8c88f14a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xxrp6\" (UID: \"941a07b0-8984-43f4-8fcd-6cdc8c88f14a\") " pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314176 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmhvp\" (UniqueName: \"kubernetes.io/projected/cee9f5f3-5282-4ec5-94a1-fd193ea66f2e-kube-api-access-gmhvp\") pod \"machine-config-operator-74547568cd-hrm6w\" (UID: \"cee9f5f3-5282-4ec5-94a1-fd193ea66f2e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hrm6w" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314308 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/66ea3a98-dc66-4baa-be63-dc53b00d3290-metrics-tls\") pod \"dns-default-md2nq\" (UID: \"66ea3a98-dc66-4baa-be63-dc53b00d3290\") " pod="openshift-dns/dns-default-md2nq" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314324 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cee9f5f3-5282-4ec5-94a1-fd193ea66f2e-proxy-tls\") pod \"machine-config-operator-74547568cd-hrm6w\" (UID: \"cee9f5f3-5282-4ec5-94a1-fd193ea66f2e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hrm6w" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314391 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/df922398-eb58-4f23-bfab-cfe9b90908c6-trusted-ca\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314409 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/e06a3c18-5c45-4153-8b28-4cf918a55970-node-bootstrap-token\") pod \"machine-config-server-5n6pv\" (UID: \"e06a3c18-5c45-4153-8b28-4cf918a55970\") " pod="openshift-machine-config-operator/machine-config-server-5n6pv" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314462 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j25rp\" (UniqueName: \"kubernetes.io/projected/b74aaf8a-71e4-4b9a-8212-c0f1cb2fd01b-kube-api-access-j25rp\") pod \"multus-admission-controller-857f4d67dd-ndhjw\" (UID: \"b74aaf8a-71e4-4b9a-8212-c0f1cb2fd01b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ndhjw" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314479 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a0c610b9-d890-49ee-b321-a3f1efba4b0a-csi-data-dir\") pod \"csi-hostpathplugin-5s9zf\" (UID: \"a0c610b9-d890-49ee-b321-a3f1efba4b0a\") " pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314515 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sklws\" (UniqueName: \"kubernetes.io/projected/585041c4-cd72-4116-b7cf-5801881f349c-kube-api-access-sklws\") pod \"machine-approver-56656f9798-2zfd9\" (UID: \"585041c4-cd72-4116-b7cf-5801881f349c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2zfd9" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314537 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/873d8371-2f45-409a-b18d-f4c7dbdabeea-tmpfs\") pod \"packageserver-d55dfcdfc-t6q48\" (UID: \"873d8371-2f45-409a-b18d-f4c7dbdabeea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314561 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2eec2888-f042-4de8-a6f7-b2bb2bb3db14-config\") pod \"service-ca-operator-777779d784-spr46\" (UID: \"2eec2888-f042-4de8-a6f7-b2bb2bb3db14\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-spr46" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314585 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6wtf\" (UniqueName: \"kubernetes.io/projected/e06a3c18-5c45-4153-8b28-4cf918a55970-kube-api-access-r6wtf\") pod \"machine-config-server-5n6pv\" (UID: \"e06a3c18-5c45-4153-8b28-4cf918a55970\") " pod="openshift-machine-config-operator/machine-config-server-5n6pv" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314630 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v6hs\" (UniqueName: \"kubernetes.io/projected/ce5a8998-bb68-4764-aa7d-a47ec23c4514-kube-api-access-5v6hs\") pod \"collect-profiles-29447010-fl8rj\" (UID: \"ce5a8998-bb68-4764-aa7d-a47ec23c4514\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314650 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9ef59357-6b1b-491e-a43d-3d436b09d4ff-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-22jvt\" (UID: \"9ef59357-6b1b-491e-a43d-3d436b09d4ff\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22jvt" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314665 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66ea3a98-dc66-4baa-be63-dc53b00d3290-config-volume\") pod \"dns-default-md2nq\" (UID: \"66ea3a98-dc66-4baa-be63-dc53b00d3290\") " pod="openshift-dns/dns-default-md2nq" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314724 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/941a07b0-8984-43f4-8fcd-6cdc8c88f14a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xxrp6\" (UID: \"941a07b0-8984-43f4-8fcd-6cdc8c88f14a\") " pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314855 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e4f56101-c810-4e1f-96e8-c7cbcb2ec719-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zchqn\" (UID: \"e4f56101-c810-4e1f-96e8-c7cbcb2ec719\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zchqn" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314899 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/df922398-eb58-4f23-bfab-cfe9b90908c6-registry-tls\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314953 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/df922398-eb58-4f23-bfab-cfe9b90908c6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314973 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/873d8371-2f45-409a-b18d-f4c7dbdabeea-webhook-cert\") pod \"packageserver-d55dfcdfc-t6q48\" (UID: \"873d8371-2f45-409a-b18d-f4c7dbdabeea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.314994 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbszq\" (UniqueName: \"kubernetes.io/projected/85b3562e-4570-4b9b-baa5-a20fe55f13ca-kube-api-access-bbszq\") pod \"package-server-manager-789f6589d5-c2tqv\" (UID: \"85b3562e-4570-4b9b-baa5-a20fe55f13ca\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-c2tqv" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.315025 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4ea539b6-ffe0-4393-af8d-62579266697c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-r6jld\" (UID: \"4ea539b6-ffe0-4393-af8d-62579266697c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.315043 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pb4cf\" (UniqueName: \"kubernetes.io/projected/66ea3a98-dc66-4baa-be63-dc53b00d3290-kube-api-access-pb4cf\") pod \"dns-default-md2nq\" (UID: \"66ea3a98-dc66-4baa-be63-dc53b00d3290\") " pod="openshift-dns/dns-default-md2nq" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.315207 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/585041c4-cd72-4116-b7cf-5801881f349c-auth-proxy-config\") pod \"machine-approver-56656f9798-2zfd9\" (UID: \"585041c4-cd72-4116-b7cf-5801881f349c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2zfd9" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.318363 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/df922398-eb58-4f23-bfab-cfe9b90908c6-trusted-ca\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.322183 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/df922398-eb58-4f23-bfab-cfe9b90908c6-ca-trust-extracted\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.322400 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9ef59357-6b1b-491e-a43d-3d436b09d4ff-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-22jvt\" (UID: \"9ef59357-6b1b-491e-a43d-3d436b09d4ff\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22jvt" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.322725 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/585041c4-cd72-4116-b7cf-5801881f349c-config\") pod \"machine-approver-56656f9798-2zfd9\" (UID: \"585041c4-cd72-4116-b7cf-5801881f349c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2zfd9" Dec 27 07:44:47 crc kubenswrapper[4934]: E1227 07:44:47.323375 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:47.823360005 +0000 UTC m=+148.643800599 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.330259 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/df922398-eb58-4f23-bfab-cfe9b90908c6-installation-pull-secrets\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.332210 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/585041c4-cd72-4116-b7cf-5801881f349c-machine-approver-tls\") pod \"machine-approver-56656f9798-2zfd9\" (UID: \"585041c4-cd72-4116-b7cf-5801881f349c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2zfd9" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.338402 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/df922398-eb58-4f23-bfab-cfe9b90908c6-registry-tls\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.340263 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9ef59357-6b1b-491e-a43d-3d436b09d4ff-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-22jvt\" (UID: \"9ef59357-6b1b-491e-a43d-3d436b09d4ff\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22jvt" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.342857 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-j5cpz"] Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.375748 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kdpx\" (UniqueName: \"kubernetes.io/projected/9ef59357-6b1b-491e-a43d-3d436b09d4ff-kube-api-access-5kdpx\") pod \"cluster-image-registry-operator-dc59b4c8b-22jvt\" (UID: \"9ef59357-6b1b-491e-a43d-3d436b09d4ff\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22jvt" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.384875 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9q69n\" (UniqueName: \"kubernetes.io/projected/df922398-eb58-4f23-bfab-cfe9b90908c6-kube-api-access-9q69n\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.391834 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-nr4vm"] Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.395679 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hn7lm"] Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.400762 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j25rp\" (UniqueName: \"kubernetes.io/projected/b74aaf8a-71e4-4b9a-8212-c0f1cb2fd01b-kube-api-access-j25rp\") pod \"multus-admission-controller-857f4d67dd-ndhjw\" (UID: \"b74aaf8a-71e4-4b9a-8212-c0f1cb2fd01b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ndhjw" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.405646 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8msj4"] Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.413262 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b74aaf8a-71e4-4b9a-8212-c0f1cb2fd01b-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ndhjw\" (UID: \"b74aaf8a-71e4-4b9a-8212-c0f1cb2fd01b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ndhjw" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.413594 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/df922398-eb58-4f23-bfab-cfe9b90908c6-registry-certificates\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.416492 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.416640 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/85b3562e-4570-4b9b-baa5-a20fe55f13ca-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-c2tqv\" (UID: \"85b3562e-4570-4b9b-baa5-a20fe55f13ca\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-c2tqv" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.416664 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ac1de940-6952-466a-be96-52c43b1a90a2-stats-auth\") pod \"router-default-5444994796-cpts5\" (UID: \"ac1de940-6952-466a-be96-52c43b1a90a2\") " pod="openshift-ingress/router-default-5444994796-cpts5" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.416701 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cee9f5f3-5282-4ec5-94a1-fd193ea66f2e-auth-proxy-config\") pod \"machine-config-operator-74547568cd-hrm6w\" (UID: \"cee9f5f3-5282-4ec5-94a1-fd193ea66f2e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hrm6w" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.416731 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/e06a3c18-5c45-4153-8b28-4cf918a55970-certs\") pod \"machine-config-server-5n6pv\" (UID: \"e06a3c18-5c45-4153-8b28-4cf918a55970\") " pod="openshift-machine-config-operator/machine-config-server-5n6pv" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.416757 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrs9l\" (UniqueName: \"kubernetes.io/projected/a0c610b9-d890-49ee-b321-a3f1efba4b0a-kube-api-access-lrs9l\") pod \"csi-hostpathplugin-5s9zf\" (UID: \"a0c610b9-d890-49ee-b321-a3f1efba4b0a\") " pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.416777 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cee9f5f3-5282-4ec5-94a1-fd193ea66f2e-images\") pod \"machine-config-operator-74547568cd-hrm6w\" (UID: \"cee9f5f3-5282-4ec5-94a1-fd193ea66f2e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hrm6w" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.416800 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z78ms\" (UniqueName: \"kubernetes.io/projected/e4f56101-c810-4e1f-96e8-c7cbcb2ec719-kube-api-access-z78ms\") pod \"control-plane-machine-set-operator-78cbb6b69f-zchqn\" (UID: \"e4f56101-c810-4e1f-96e8-c7cbcb2ec719\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zchqn" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.416841 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmhvp\" (UniqueName: \"kubernetes.io/projected/cee9f5f3-5282-4ec5-94a1-fd193ea66f2e-kube-api-access-gmhvp\") pod \"machine-config-operator-74547568cd-hrm6w\" (UID: \"cee9f5f3-5282-4ec5-94a1-fd193ea66f2e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hrm6w" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.416868 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/941a07b0-8984-43f4-8fcd-6cdc8c88f14a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xxrp6\" (UID: \"941a07b0-8984-43f4-8fcd-6cdc8c88f14a\") " pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.417424 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cee9f5f3-5282-4ec5-94a1-fd193ea66f2e-proxy-tls\") pod \"machine-config-operator-74547568cd-hrm6w\" (UID: \"cee9f5f3-5282-4ec5-94a1-fd193ea66f2e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hrm6w" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.417553 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/66ea3a98-dc66-4baa-be63-dc53b00d3290-metrics-tls\") pod \"dns-default-md2nq\" (UID: \"66ea3a98-dc66-4baa-be63-dc53b00d3290\") " pod="openshift-dns/dns-default-md2nq" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.417582 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/e06a3c18-5c45-4153-8b28-4cf918a55970-node-bootstrap-token\") pod \"machine-config-server-5n6pv\" (UID: \"e06a3c18-5c45-4153-8b28-4cf918a55970\") " pod="openshift-machine-config-operator/machine-config-server-5n6pv" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.417607 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a0c610b9-d890-49ee-b321-a3f1efba4b0a-csi-data-dir\") pod \"csi-hostpathplugin-5s9zf\" (UID: \"a0c610b9-d890-49ee-b321-a3f1efba4b0a\") " pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.417639 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/873d8371-2f45-409a-b18d-f4c7dbdabeea-tmpfs\") pod \"packageserver-d55dfcdfc-t6q48\" (UID: \"873d8371-2f45-409a-b18d-f4c7dbdabeea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.417659 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2eec2888-f042-4de8-a6f7-b2bb2bb3db14-config\") pod \"service-ca-operator-777779d784-spr46\" (UID: \"2eec2888-f042-4de8-a6f7-b2bb2bb3db14\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-spr46" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.417678 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6wtf\" (UniqueName: \"kubernetes.io/projected/e06a3c18-5c45-4153-8b28-4cf918a55970-kube-api-access-r6wtf\") pod \"machine-config-server-5n6pv\" (UID: \"e06a3c18-5c45-4153-8b28-4cf918a55970\") " pod="openshift-machine-config-operator/machine-config-server-5n6pv" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.417698 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66ea3a98-dc66-4baa-be63-dc53b00d3290-config-volume\") pod \"dns-default-md2nq\" (UID: \"66ea3a98-dc66-4baa-be63-dc53b00d3290\") " pod="openshift-dns/dns-default-md2nq" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.417718 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v6hs\" (UniqueName: \"kubernetes.io/projected/ce5a8998-bb68-4764-aa7d-a47ec23c4514-kube-api-access-5v6hs\") pod \"collect-profiles-29447010-fl8rj\" (UID: \"ce5a8998-bb68-4764-aa7d-a47ec23c4514\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.417741 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/941a07b0-8984-43f4-8fcd-6cdc8c88f14a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xxrp6\" (UID: \"941a07b0-8984-43f4-8fcd-6cdc8c88f14a\") " pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.417770 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e4f56101-c810-4e1f-96e8-c7cbcb2ec719-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zchqn\" (UID: \"e4f56101-c810-4e1f-96e8-c7cbcb2ec719\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zchqn" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.417795 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/873d8371-2f45-409a-b18d-f4c7dbdabeea-webhook-cert\") pod \"packageserver-d55dfcdfc-t6q48\" (UID: \"873d8371-2f45-409a-b18d-f4c7dbdabeea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.417820 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbszq\" (UniqueName: \"kubernetes.io/projected/85b3562e-4570-4b9b-baa5-a20fe55f13ca-kube-api-access-bbszq\") pod \"package-server-manager-789f6589d5-c2tqv\" (UID: \"85b3562e-4570-4b9b-baa5-a20fe55f13ca\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-c2tqv" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.417843 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pb4cf\" (UniqueName: \"kubernetes.io/projected/66ea3a98-dc66-4baa-be63-dc53b00d3290-kube-api-access-pb4cf\") pod \"dns-default-md2nq\" (UID: \"66ea3a98-dc66-4baa-be63-dc53b00d3290\") " pod="openshift-dns/dns-default-md2nq" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.417864 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4ea539b6-ffe0-4393-af8d-62579266697c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-r6jld\" (UID: \"4ea539b6-ffe0-4393-af8d-62579266697c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.417889 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/873d8371-2f45-409a-b18d-f4c7dbdabeea-apiservice-cert\") pod \"packageserver-d55dfcdfc-t6q48\" (UID: \"873d8371-2f45-409a-b18d-f4c7dbdabeea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.418117 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfmw2\" (UniqueName: \"kubernetes.io/projected/9552ddff-a57b-4311-a6d5-552a8b8efe19-kube-api-access-xfmw2\") pod \"ingress-canary-7tf92\" (UID: \"9552ddff-a57b-4311-a6d5-552a8b8efe19\") " pod="openshift-ingress-canary/ingress-canary-7tf92" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.418149 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4ea539b6-ffe0-4393-af8d-62579266697c-srv-cert\") pod \"olm-operator-6b444d44fb-r6jld\" (UID: \"4ea539b6-ffe0-4393-af8d-62579266697c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.418175 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ac1de940-6952-466a-be96-52c43b1a90a2-default-certificate\") pod \"router-default-5444994796-cpts5\" (UID: \"ac1de940-6952-466a-be96-52c43b1a90a2\") " pod="openshift-ingress/router-default-5444994796-cpts5" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.418196 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a0c610b9-d890-49ee-b321-a3f1efba4b0a-plugins-dir\") pod \"csi-hostpathplugin-5s9zf\" (UID: \"a0c610b9-d890-49ee-b321-a3f1efba4b0a\") " pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.418217 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jctwn\" (UniqueName: \"kubernetes.io/projected/4ea539b6-ffe0-4393-af8d-62579266697c-kube-api-access-jctwn\") pod \"olm-operator-6b444d44fb-r6jld\" (UID: \"4ea539b6-ffe0-4393-af8d-62579266697c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.418240 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a0c610b9-d890-49ee-b321-a3f1efba4b0a-registration-dir\") pod \"csi-hostpathplugin-5s9zf\" (UID: \"a0c610b9-d890-49ee-b321-a3f1efba4b0a\") " pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.418262 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnzjz\" (UniqueName: \"kubernetes.io/projected/873d8371-2f45-409a-b18d-f4c7dbdabeea-kube-api-access-gnzjz\") pod \"packageserver-d55dfcdfc-t6q48\" (UID: \"873d8371-2f45-409a-b18d-f4c7dbdabeea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.418291 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce5a8998-bb68-4764-aa7d-a47ec23c4514-config-volume\") pod \"collect-profiles-29447010-fl8rj\" (UID: \"ce5a8998-bb68-4764-aa7d-a47ec23c4514\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.418311 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc6ns\" (UniqueName: \"kubernetes.io/projected/ac1de940-6952-466a-be96-52c43b1a90a2-kube-api-access-mc6ns\") pod \"router-default-5444994796-cpts5\" (UID: \"ac1de940-6952-466a-be96-52c43b1a90a2\") " pod="openshift-ingress/router-default-5444994796-cpts5" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.418331 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce5a8998-bb68-4764-aa7d-a47ec23c4514-secret-volume\") pod \"collect-profiles-29447010-fl8rj\" (UID: \"ce5a8998-bb68-4764-aa7d-a47ec23c4514\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.418349 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9552ddff-a57b-4311-a6d5-552a8b8efe19-cert\") pod \"ingress-canary-7tf92\" (UID: \"9552ddff-a57b-4311-a6d5-552a8b8efe19\") " pod="openshift-ingress-canary/ingress-canary-7tf92" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.418375 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2eec2888-f042-4de8-a6f7-b2bb2bb3db14-serving-cert\") pod \"service-ca-operator-777779d784-spr46\" (UID: \"2eec2888-f042-4de8-a6f7-b2bb2bb3db14\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-spr46" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.418396 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ac1de940-6952-466a-be96-52c43b1a90a2-metrics-certs\") pod \"router-default-5444994796-cpts5\" (UID: \"ac1de940-6952-466a-be96-52c43b1a90a2\") " pod="openshift-ingress/router-default-5444994796-cpts5" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.418422 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftz4h\" (UniqueName: \"kubernetes.io/projected/941a07b0-8984-43f4-8fcd-6cdc8c88f14a-kube-api-access-ftz4h\") pod \"marketplace-operator-79b997595-xxrp6\" (UID: \"941a07b0-8984-43f4-8fcd-6cdc8c88f14a\") " pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.419179 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a0c610b9-d890-49ee-b321-a3f1efba4b0a-mountpoint-dir\") pod \"csi-hostpathplugin-5s9zf\" (UID: \"a0c610b9-d890-49ee-b321-a3f1efba4b0a\") " pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.419206 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a0c610b9-d890-49ee-b321-a3f1efba4b0a-socket-dir\") pod \"csi-hostpathplugin-5s9zf\" (UID: \"a0c610b9-d890-49ee-b321-a3f1efba4b0a\") " pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.419233 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac1de940-6952-466a-be96-52c43b1a90a2-service-ca-bundle\") pod \"router-default-5444994796-cpts5\" (UID: \"ac1de940-6952-466a-be96-52c43b1a90a2\") " pod="openshift-ingress/router-default-5444994796-cpts5" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.419260 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26nvp\" (UniqueName: \"kubernetes.io/projected/2eec2888-f042-4de8-a6f7-b2bb2bb3db14-kube-api-access-26nvp\") pod \"service-ca-operator-777779d784-spr46\" (UID: \"2eec2888-f042-4de8-a6f7-b2bb2bb3db14\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-spr46" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.421830 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cee9f5f3-5282-4ec5-94a1-fd193ea66f2e-images\") pod \"machine-config-operator-74547568cd-hrm6w\" (UID: \"cee9f5f3-5282-4ec5-94a1-fd193ea66f2e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hrm6w" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.423258 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/e06a3c18-5c45-4153-8b28-4cf918a55970-certs\") pod \"machine-config-server-5n6pv\" (UID: \"e06a3c18-5c45-4153-8b28-4cf918a55970\") " pod="openshift-machine-config-operator/machine-config-server-5n6pv" Dec 27 07:44:47 crc kubenswrapper[4934]: E1227 07:44:47.423344 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:47.923324853 +0000 UTC m=+148.743765447 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.431407 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/873d8371-2f45-409a-b18d-f4c7dbdabeea-tmpfs\") pod \"packageserver-d55dfcdfc-t6q48\" (UID: \"873d8371-2f45-409a-b18d-f4c7dbdabeea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.433233 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/941a07b0-8984-43f4-8fcd-6cdc8c88f14a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xxrp6\" (UID: \"941a07b0-8984-43f4-8fcd-6cdc8c88f14a\") " pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.436860 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ac1de940-6952-466a-be96-52c43b1a90a2-stats-auth\") pod \"router-default-5444994796-cpts5\" (UID: \"ac1de940-6952-466a-be96-52c43b1a90a2\") " pod="openshift-ingress/router-default-5444994796-cpts5" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.439387 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e4f56101-c810-4e1f-96e8-c7cbcb2ec719-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zchqn\" (UID: \"e4f56101-c810-4e1f-96e8-c7cbcb2ec719\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zchqn" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.443507 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/85b3562e-4570-4b9b-baa5-a20fe55f13ca-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-c2tqv\" (UID: \"85b3562e-4570-4b9b-baa5-a20fe55f13ca\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-c2tqv" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.443945 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cee9f5f3-5282-4ec5-94a1-fd193ea66f2e-auth-proxy-config\") pod \"machine-config-operator-74547568cd-hrm6w\" (UID: \"cee9f5f3-5282-4ec5-94a1-fd193ea66f2e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hrm6w" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.449398 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4ea539b6-ffe0-4393-af8d-62579266697c-srv-cert\") pod \"olm-operator-6b444d44fb-r6jld\" (UID: \"4ea539b6-ffe0-4393-af8d-62579266697c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.451947 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ac1de940-6952-466a-be96-52c43b1a90a2-default-certificate\") pod \"router-default-5444994796-cpts5\" (UID: \"ac1de940-6952-466a-be96-52c43b1a90a2\") " pod="openshift-ingress/router-default-5444994796-cpts5" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.452232 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a0c610b9-d890-49ee-b321-a3f1efba4b0a-plugins-dir\") pod \"csi-hostpathplugin-5s9zf\" (UID: \"a0c610b9-d890-49ee-b321-a3f1efba4b0a\") " pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.452371 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a0c610b9-d890-49ee-b321-a3f1efba4b0a-registration-dir\") pod \"csi-hostpathplugin-5s9zf\" (UID: \"a0c610b9-d890-49ee-b321-a3f1efba4b0a\") " pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.453172 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a0c610b9-d890-49ee-b321-a3f1efba4b0a-mountpoint-dir\") pod \"csi-hostpathplugin-5s9zf\" (UID: \"a0c610b9-d890-49ee-b321-a3f1efba4b0a\") " pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.453665 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/873d8371-2f45-409a-b18d-f4c7dbdabeea-webhook-cert\") pod \"packageserver-d55dfcdfc-t6q48\" (UID: \"873d8371-2f45-409a-b18d-f4c7dbdabeea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.456739 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9552ddff-a57b-4311-a6d5-552a8b8efe19-cert\") pod \"ingress-canary-7tf92\" (UID: \"9552ddff-a57b-4311-a6d5-552a8b8efe19\") " pod="openshift-ingress-canary/ingress-canary-7tf92" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.456930 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a0c610b9-d890-49ee-b321-a3f1efba4b0a-socket-dir\") pod \"csi-hostpathplugin-5s9zf\" (UID: \"a0c610b9-d890-49ee-b321-a3f1efba4b0a\") " pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.458453 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sklws\" (UniqueName: \"kubernetes.io/projected/585041c4-cd72-4116-b7cf-5801881f349c-kube-api-access-sklws\") pod \"machine-approver-56656f9798-2zfd9\" (UID: \"585041c4-cd72-4116-b7cf-5801881f349c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2zfd9" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.458826 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cee9f5f3-5282-4ec5-94a1-fd193ea66f2e-proxy-tls\") pod \"machine-config-operator-74547568cd-hrm6w\" (UID: \"cee9f5f3-5282-4ec5-94a1-fd193ea66f2e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hrm6w" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.459119 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce5a8998-bb68-4764-aa7d-a47ec23c4514-secret-volume\") pod \"collect-profiles-29447010-fl8rj\" (UID: \"ce5a8998-bb68-4764-aa7d-a47ec23c4514\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.459540 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/941a07b0-8984-43f4-8fcd-6cdc8c88f14a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xxrp6\" (UID: \"941a07b0-8984-43f4-8fcd-6cdc8c88f14a\") " pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.459660 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/66ea3a98-dc66-4baa-be63-dc53b00d3290-metrics-tls\") pod \"dns-default-md2nq\" (UID: \"66ea3a98-dc66-4baa-be63-dc53b00d3290\") " pod="openshift-dns/dns-default-md2nq" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.459702 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/e06a3c18-5c45-4153-8b28-4cf918a55970-node-bootstrap-token\") pod \"machine-config-server-5n6pv\" (UID: \"e06a3c18-5c45-4153-8b28-4cf918a55970\") " pod="openshift-machine-config-operator/machine-config-server-5n6pv" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.459802 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a0c610b9-d890-49ee-b321-a3f1efba4b0a-csi-data-dir\") pod \"csi-hostpathplugin-5s9zf\" (UID: \"a0c610b9-d890-49ee-b321-a3f1efba4b0a\") " pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.461345 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2zfd9" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.463616 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4ea539b6-ffe0-4393-af8d-62579266697c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-r6jld\" (UID: \"4ea539b6-ffe0-4393-af8d-62579266697c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.466746 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ac1de940-6952-466a-be96-52c43b1a90a2-metrics-certs\") pod \"router-default-5444994796-cpts5\" (UID: \"ac1de940-6952-466a-be96-52c43b1a90a2\") " pod="openshift-ingress/router-default-5444994796-cpts5" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.467306 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/df922398-eb58-4f23-bfab-cfe9b90908c6-bound-sa-token\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.472414 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/873d8371-2f45-409a-b18d-f4c7dbdabeea-apiservice-cert\") pod \"packageserver-d55dfcdfc-t6q48\" (UID: \"873d8371-2f45-409a-b18d-f4c7dbdabeea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.472859 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9ef59357-6b1b-491e-a43d-3d436b09d4ff-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-22jvt\" (UID: \"9ef59357-6b1b-491e-a43d-3d436b09d4ff\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22jvt" Dec 27 07:44:47 crc kubenswrapper[4934]: W1227 07:44:47.484411 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ebeee10_d90f_4672_8bfa_5f10fa929d31.slice/crio-08640a6b54dd159d50ddb7fb9bf35c5edb74292834b4709a8b556ab64a39e781 WatchSource:0}: Error finding container 08640a6b54dd159d50ddb7fb9bf35c5edb74292834b4709a8b556ab64a39e781: Status 404 returned error can't find the container with id 08640a6b54dd159d50ddb7fb9bf35c5edb74292834b4709a8b556ab64a39e781 Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.517993 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrs9l\" (UniqueName: \"kubernetes.io/projected/a0c610b9-d890-49ee-b321-a3f1efba4b0a-kube-api-access-lrs9l\") pod \"csi-hostpathplugin-5s9zf\" (UID: \"a0c610b9-d890-49ee-b321-a3f1efba4b0a\") " pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.523584 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: E1227 07:44:47.524974 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:48.02494105 +0000 UTC m=+148.845381644 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.525736 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z78ms\" (UniqueName: \"kubernetes.io/projected/e4f56101-c810-4e1f-96e8-c7cbcb2ec719-kube-api-access-z78ms\") pod \"control-plane-machine-set-operator-78cbb6b69f-zchqn\" (UID: \"e4f56101-c810-4e1f-96e8-c7cbcb2ec719\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zchqn" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.534875 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66ea3a98-dc66-4baa-be63-dc53b00d3290-config-volume\") pod \"dns-default-md2nq\" (UID: \"66ea3a98-dc66-4baa-be63-dc53b00d3290\") " pod="openshift-dns/dns-default-md2nq" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.535066 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2eec2888-f042-4de8-a6f7-b2bb2bb3db14-config\") pod \"service-ca-operator-777779d784-spr46\" (UID: \"2eec2888-f042-4de8-a6f7-b2bb2bb3db14\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-spr46" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.536546 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce5a8998-bb68-4764-aa7d-a47ec23c4514-config-volume\") pod \"collect-profiles-29447010-fl8rj\" (UID: \"ce5a8998-bb68-4764-aa7d-a47ec23c4514\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.539092 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac1de940-6952-466a-be96-52c43b1a90a2-service-ca-bundle\") pod \"router-default-5444994796-cpts5\" (UID: \"ac1de940-6952-466a-be96-52c43b1a90a2\") " pod="openshift-ingress/router-default-5444994796-cpts5" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.543519 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2eec2888-f042-4de8-a6f7-b2bb2bb3db14-serving-cert\") pod \"service-ca-operator-777779d784-spr46\" (UID: \"2eec2888-f042-4de8-a6f7-b2bb2bb3db14\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-spr46" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.545469 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ndhjw" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.546681 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmhvp\" (UniqueName: \"kubernetes.io/projected/cee9f5f3-5282-4ec5-94a1-fd193ea66f2e-kube-api-access-gmhvp\") pod \"machine-config-operator-74547568cd-hrm6w\" (UID: \"cee9f5f3-5282-4ec5-94a1-fd193ea66f2e\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hrm6w" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.562547 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfmw2\" (UniqueName: \"kubernetes.io/projected/9552ddff-a57b-4311-a6d5-552a8b8efe19-kube-api-access-xfmw2\") pod \"ingress-canary-7tf92\" (UID: \"9552ddff-a57b-4311-a6d5-552a8b8efe19\") " pod="openshift-ingress-canary/ingress-canary-7tf92" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.574799 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-8s8bp"] Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.582445 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v6hs\" (UniqueName: \"kubernetes.io/projected/ce5a8998-bb68-4764-aa7d-a47ec23c4514-kube-api-access-5v6hs\") pod \"collect-profiles-29447010-fl8rj\" (UID: \"ce5a8998-bb68-4764-aa7d-a47ec23c4514\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.591036 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws"] Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.594884 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6bmfr"] Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.599071 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jctwn\" (UniqueName: \"kubernetes.io/projected/4ea539b6-ffe0-4393-af8d-62579266697c-kube-api-access-jctwn\") pod \"olm-operator-6b444d44fb-r6jld\" (UID: \"4ea539b6-ffe0-4393-af8d-62579266697c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.619718 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnzjz\" (UniqueName: \"kubernetes.io/projected/873d8371-2f45-409a-b18d-f4c7dbdabeea-kube-api-access-gnzjz\") pod \"packageserver-d55dfcdfc-t6q48\" (UID: \"873d8371-2f45-409a-b18d-f4c7dbdabeea\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.628726 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:47 crc kubenswrapper[4934]: E1227 07:44:47.628922 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:48.128870345 +0000 UTC m=+148.949310949 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.629356 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: E1227 07:44:47.629738 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:48.129727181 +0000 UTC m=+148.950167775 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.635727 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hrm6w" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.648335 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc6ns\" (UniqueName: \"kubernetes.io/projected/ac1de940-6952-466a-be96-52c43b1a90a2-kube-api-access-mc6ns\") pod \"router-default-5444994796-cpts5\" (UID: \"ac1de940-6952-466a-be96-52c43b1a90a2\") " pod="openshift-ingress/router-default-5444994796-cpts5" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.650777 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zchqn" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.666037 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.673167 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.679162 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-cpts5" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.682797 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbszq\" (UniqueName: \"kubernetes.io/projected/85b3562e-4570-4b9b-baa5-a20fe55f13ca-kube-api-access-bbszq\") pod \"package-server-manager-789f6589d5-c2tqv\" (UID: \"85b3562e-4570-4b9b-baa5-a20fe55f13ca\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-c2tqv" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.687690 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.691450 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-z76rm"] Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.693723 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-7tf92" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.701965 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26nvp\" (UniqueName: \"kubernetes.io/projected/2eec2888-f042-4de8-a6f7-b2bb2bb3db14-kube-api-access-26nvp\") pod \"service-ca-operator-777779d784-spr46\" (UID: \"2eec2888-f042-4de8-a6f7-b2bb2bb3db14\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-spr46" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.711782 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ph956"] Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.714395 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-wvjnl"] Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.720040 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-jckgt"] Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.720141 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.723034 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6wtf\" (UniqueName: \"kubernetes.io/projected/e06a3c18-5c45-4153-8b28-4cf918a55970-kube-api-access-r6wtf\") pod \"machine-config-server-5n6pv\" (UID: \"e06a3c18-5c45-4153-8b28-4cf918a55970\") " pod="openshift-machine-config-operator/machine-config-server-5n6pv" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.726810 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22jvt" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.729645 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p"] Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.729942 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:47 crc kubenswrapper[4934]: E1227 07:44:47.730306 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:48.230285956 +0000 UTC m=+149.050726550 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.739286 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-5n6pv" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.755654 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pb4cf\" (UniqueName: \"kubernetes.io/projected/66ea3a98-dc66-4baa-be63-dc53b00d3290-kube-api-access-pb4cf\") pod \"dns-default-md2nq\" (UID: \"66ea3a98-dc66-4baa-be63-dc53b00d3290\") " pod="openshift-dns/dns-default-md2nq" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.758820 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftz4h\" (UniqueName: \"kubernetes.io/projected/941a07b0-8984-43f4-8fcd-6cdc8c88f14a-kube-api-access-ftz4h\") pod \"marketplace-operator-79b997595-xxrp6\" (UID: \"941a07b0-8984-43f4-8fcd-6cdc8c88f14a\") " pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.793980 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zcl7d"] Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.807897 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hh662"] Dec 27 07:44:47 crc kubenswrapper[4934]: W1227 07:44:47.810525 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod753c599b_a401_481e_a5b3_e2dd6ed2c4e7.slice/crio-5a245c846367560e4692306fde5c91f0053c28af5b4e76634f21644b5eccc034 WatchSource:0}: Error finding container 5a245c846367560e4692306fde5c91f0053c28af5b4e76634f21644b5eccc034: Status 404 returned error can't find the container with id 5a245c846367560e4692306fde5c91f0053c28af5b4e76634f21644b5eccc034 Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.831093 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-whhqx"] Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.831150 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gdpgt"] Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.831793 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.831800 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-mn55p"] Dec 27 07:44:47 crc kubenswrapper[4934]: E1227 07:44:47.832097 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:48.332068938 +0000 UTC m=+149.152509532 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.832941 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7"] Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.836484 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8npds"] Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.924666 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.928359 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-c2tqv" Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.932321 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.933723 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-w4pmg"] Dec 27 07:44:47 crc kubenswrapper[4934]: E1227 07:44:47.933737 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:48.433713926 +0000 UTC m=+149.254154520 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:47 crc kubenswrapper[4934]: I1227 07:44:47.950251 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-spr46" Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.002859 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ndhjw"] Dec 27 07:44:48 crc kubenswrapper[4934]: W1227 07:44:48.030996 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd2142a6_e9cb_432d_adf9_61b5d3f3c5b1.slice/crio-caaaf68d7296c7d4f8eaeacaa6f1ddccdc549262f02122867f7ec9accf0bbcc3 WatchSource:0}: Error finding container caaaf68d7296c7d4f8eaeacaa6f1ddccdc549262f02122867f7ec9accf0bbcc3: Status 404 returned error can't find the container with id caaaf68d7296c7d4f8eaeacaa6f1ddccdc549262f02122867f7ec9accf0bbcc3 Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.031948 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-md2nq" Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.035986 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:48 crc kubenswrapper[4934]: E1227 07:44:48.037407 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:48.537392934 +0000 UTC m=+149.357833528 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.103780 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-hrm6w"] Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.139628 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:48 crc kubenswrapper[4934]: E1227 07:44:48.139852 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:48.639818144 +0000 UTC m=+149.460258738 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.139893 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:48 crc kubenswrapper[4934]: E1227 07:44:48.140337 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:48.640321939 +0000 UTC m=+149.460762533 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:48 crc kubenswrapper[4934]: W1227 07:44:48.142226 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb74aaf8a_71e4_4b9a_8212_c0f1cb2fd01b.slice/crio-bc307579136cc26e17245e2595315e7827567474730baca38c34512985a960e0 WatchSource:0}: Error finding container bc307579136cc26e17245e2595315e7827567474730baca38c34512985a960e0: Status 404 returned error can't find the container with id bc307579136cc26e17245e2595315e7827567474730baca38c34512985a960e0 Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.241049 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:48 crc kubenswrapper[4934]: E1227 07:44:48.241270 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:48.741243316 +0000 UTC m=+149.561683910 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.241903 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:48 crc kubenswrapper[4934]: E1227 07:44:48.242159 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:48.742151303 +0000 UTC m=+149.562591897 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.312041 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" event={"ID":"901df0a0-1dae-41aa-8d29-0981524d9312","Type":"ContainerStarted","Data":"168d3ccd34a0cb53256bd43f1c7d232e7863a1a4b89e34bb912dd372159ecab7"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.321719 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wvjnl" event={"ID":"cb832863-8ab8-459a-8d73-3c464b8caf34","Type":"ContainerStarted","Data":"4a0056af32bdeea097780cfcf5e966f0dbd6386da060613529a734332ecb8a7b"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.345507 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:48 crc kubenswrapper[4934]: E1227 07:44:48.345821 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:48.84580614 +0000 UTC m=+149.666246734 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.350977 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" event={"ID":"145ee02a-c7aa-4990-a288-e6f69f8ab030","Type":"ContainerStarted","Data":"c6ee808aeb4966fd838294c6f98112f343cb101b22e447000b13b4e01988bd73"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.372126 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" event={"ID":"5548c900-e014-4c35-b62e-455cd82f5887","Type":"ContainerStarted","Data":"fa0e38b01db058966e64f388fec75289e82c54449b95f88ee4b3b3c16f2a8db3"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.378673 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" event={"ID":"c2f61301-d046-4651-b76c-9939b18fb792","Type":"ContainerStarted","Data":"e7fc2fe75e37d007bed71c48bbf23cadc6115308f46af37e1bcd5966c0e7037c"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.378712 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" event={"ID":"c2f61301-d046-4651-b76c-9939b18fb792","Type":"ContainerStarted","Data":"fed642fe85c2daef00a5d53c79d9c70728e906b17d0c52170df1ec440f20b6ba"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.378908 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.401356 4934 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-j5cpz container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.8:6443/healthz\": dial tcp 10.217.0.8:6443: connect: connection refused" start-of-body= Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.401408 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" podUID="c2f61301-d046-4651-b76c-9939b18fb792" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.8:6443/healthz\": dial tcp 10.217.0.8:6443: connect: connection refused" Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.429272 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld"] Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.447032 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:48 crc kubenswrapper[4934]: E1227 07:44:48.448404 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:48.948392215 +0000 UTC m=+149.768832799 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.454716 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9s2ls" event={"ID":"d6342508-1473-49c4-a78f-b4e6ed9c68b0","Type":"ContainerStarted","Data":"ab6fe899781d69ae40d2cc98d49c3c0f95af69a44a565a81f02574c30995b5b4"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.467724 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zcl7d" event={"ID":"a1e4154b-2230-4499-b645-4903ea9593f1","Type":"ContainerStarted","Data":"83cb6162e3d19682b105c000745bebf2a30432d22c7430ecff1be72ef7ce20fb"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.469795 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" event={"ID":"af023fee-eae1-427a-8d70-487588422316","Type":"ContainerStarted","Data":"f60d8867faf70d855bf4d3de99d3499deb657ee9ea0285565ed1ff36290fb267"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.469821 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" event={"ID":"af023fee-eae1-427a-8d70-487588422316","Type":"ContainerStarted","Data":"4ddc5b1ec74b0ea4de89221331d22a83fd2d84941b6da2aa9ff747736f06fcb4"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.476508 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.481440 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" event={"ID":"959b5ca2-d235-4457-98ec-fe077dfcdc6f","Type":"ContainerStarted","Data":"b6f570194919b76e785ecb947ba89694a8df2a47ab393bdaaf2e4cf367a713c1"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.481473 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" event={"ID":"959b5ca2-d235-4457-98ec-fe077dfcdc6f","Type":"ContainerStarted","Data":"151cdaebfd40baaacba6cfe6685f366489e0026332eff9399b0b32cff2895e29"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.483475 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.489013 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hrm6w" event={"ID":"cee9f5f3-5282-4ec5-94a1-fd193ea66f2e","Type":"ContainerStarted","Data":"97264abc60af9b19b9da51f55cdd5721435be512801c0afdc8e8d63393af51ff"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.499663 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zchqn"] Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.502554 4934 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-5cc9p container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.502600 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" podUID="af023fee-eae1-427a-8d70-487588422316" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.508789 4934 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-r5zxc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.508883 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" podUID="959b5ca2-d235-4457-98ec-fe077dfcdc6f" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.511526 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8msj4" event={"ID":"a0d34185-bf08-456c-a5fc-3f0852d35a91","Type":"ContainerStarted","Data":"bbb8c8c732b24041df4c223b3b91a7741cd1534a879275067b7b9c3d51a9deb7"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.511573 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8msj4" event={"ID":"a0d34185-bf08-456c-a5fc-3f0852d35a91","Type":"ContainerStarted","Data":"4657f1e56dd515c64282f517f5ae8571a2de93d5c63751061f966197df002239"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.512166 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-8msj4" Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.514591 4934 patch_prober.go:28] interesting pod/downloads-7954f5f757-8msj4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.514657 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8msj4" podUID="a0d34185-bf08-456c-a5fc-3f0852d35a91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.517868 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hn7lm" event={"ID":"7ebeee10-d90f-4672-8bfa-5f10fa929d31","Type":"ContainerStarted","Data":"a81ffea13db70e9878ff2111524d105d6a54e7c340663e9491b8cba7ae101d0f"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.517909 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hn7lm" event={"ID":"7ebeee10-d90f-4672-8bfa-5f10fa929d31","Type":"ContainerStarted","Data":"08640a6b54dd159d50ddb7fb9bf35c5edb74292834b4709a8b556ab64a39e781"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.529498 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" event={"ID":"d812eb7a-ce30-4759-a558-fff3738068b6","Type":"ContainerStarted","Data":"6842504a1cdfff4637c14774c5ed92e15efb9eff144fb742e44a9d56df4e76f2"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.540277 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-z76rm" event={"ID":"12f53d40-4cdb-41e5-b620-c85db6cb410a","Type":"ContainerStarted","Data":"1c1322276e9c2eeed223d2c89d8c15d542110673ee39b6ffa430e6a5bc5d0020"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.542781 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-cpts5" event={"ID":"ac1de940-6952-466a-be96-52c43b1a90a2","Type":"ContainerStarted","Data":"6cb87c3ba51758cbbbf3c9f87ba2374b17bdb92e39567db856156d0bd1e24b6a"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.548300 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:48 crc kubenswrapper[4934]: E1227 07:44:48.548448 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:49.048421175 +0000 UTC m=+149.868861779 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.548633 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:48 crc kubenswrapper[4934]: E1227 07:44:48.552092 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:49.052063333 +0000 UTC m=+149.872504017 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.552505 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hh662" event={"ID":"0b1dac22-8539-4f61-ac28-2f420c845a2a","Type":"ContainerStarted","Data":"c18a44f65f43180baa2803b27034acbbcb3a32fac175b68467a3b0e1c59c77f2"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.558676 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gdpgt" event={"ID":"d3ba307a-a762-4e65-aee8-d30964fadea9","Type":"ContainerStarted","Data":"65108b83e8552ff180299408ebf4ee9384079d8f774e2b23611585cf823cc870"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.562986 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2zfd9" event={"ID":"585041c4-cd72-4116-b7cf-5801881f349c","Type":"ContainerStarted","Data":"96f0fd87158146f4eacfbc11760a362160903115f10ea51e0c550816efdb6ada"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.577625 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ph956" event={"ID":"c16c1071-8350-4dda-b34b-183cad108e19","Type":"ContainerStarted","Data":"b30c179826eabffb3c8006bae9a1f8733da7d96f93d3c847dc2b5cd8294f58c9"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.584848 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn55p" event={"ID":"b2a55db0-a3bd-413f-a52f-75b36654377c","Type":"ContainerStarted","Data":"2f7284662a3dbfc9ea71fffa0cf3f63bf57674f5f6f16df4314aac5126d2cc52"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.589174 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" event={"ID":"4c71cabc-c13f-4ae8-a704-10efee323e35","Type":"ContainerStarted","Data":"83375bbba20f8f09565eb5f1f87a93e99fca2453a99f9f65d2e583b23f34474d"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.591250 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-jckgt" event={"ID":"753c599b-a401-481e-a5b3-e2dd6ed2c4e7","Type":"ContainerStarted","Data":"5a245c846367560e4692306fde5c91f0053c28af5b4e76634f21644b5eccc034"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.601486 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-p8c2x" event={"ID":"091c02c2-ae22-44e8-bf6b-64b69d6d9d83","Type":"ContainerStarted","Data":"109e74a00ca3e09f9af2206b0495f03664b3ca731e65fe55e0b9e8066f0d23a3"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.642908 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8jpzb" event={"ID":"ba604131-ba14-4754-8c5e-3e4fdccdfe64","Type":"ContainerStarted","Data":"bbf9467bf19091362471d4e1378e4820a837c2f19e3075ab57f24720f7c1a273"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.649978 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:48 crc kubenswrapper[4934]: E1227 07:44:48.651099 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:49.151069413 +0000 UTC m=+149.971509997 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.666027 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ndhjw" event={"ID":"b74aaf8a-71e4-4b9a-8212-c0f1cb2fd01b","Type":"ContainerStarted","Data":"bc307579136cc26e17245e2595315e7827567474730baca38c34512985a960e0"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.733837 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bmfr" event={"ID":"3da2af1b-ff3c-420c-af84-633eaddc9368","Type":"ContainerStarted","Data":"f955f637a3acd54b029d319e0dfc1e6e9babf6fcb340eb3d0e56a56678c782d7"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.825844 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:48 crc kubenswrapper[4934]: E1227 07:44:48.830773 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:49.330754432 +0000 UTC m=+150.151195026 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.848148 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" podStartSLOduration=129.848133395 podStartE2EDuration="2m9.848133395s" podCreationTimestamp="2025-12-27 07:42:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:48.845557139 +0000 UTC m=+149.665997723" watchObservedRunningTime="2025-12-27 07:44:48.848133395 +0000 UTC m=+149.668573989" Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.865495 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" event={"ID":"5c9fc52e-279a-44a8-a779-17966e320a93","Type":"ContainerStarted","Data":"10a94a0bbdc91fc909ec077ee954e18c922833a4fcc2bdccd0151697e8430b26"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.898837 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" podStartSLOduration=130.898815169 podStartE2EDuration="2m10.898815169s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:48.894923825 +0000 UTC m=+149.715364419" watchObservedRunningTime="2025-12-27 07:44:48.898815169 +0000 UTC m=+149.719255753" Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.919436 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-whhqx" event={"ID":"46527d26-eb38-494f-85f7-f5f56b8dfada","Type":"ContainerStarted","Data":"2d2dda045bd671f551e7454b14957f64f450ab7245599c0ec3b8141503c3e74c"} Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.930429 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:48 crc kubenswrapper[4934]: E1227 07:44:48.930830 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:49.43070486 +0000 UTC m=+150.251145454 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.931066 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:48 crc kubenswrapper[4934]: E1227 07:44:48.936877 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:49.432435611 +0000 UTC m=+150.252876405 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:48 crc kubenswrapper[4934]: I1227 07:44:48.941720 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hn7lm" podStartSLOduration=130.941698004 podStartE2EDuration="2m10.941698004s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:48.941304643 +0000 UTC m=+149.761745247" watchObservedRunningTime="2025-12-27 07:44:48.941698004 +0000 UTC m=+149.762138598" Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.003621 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" podStartSLOduration=131.00359869 podStartE2EDuration="2m11.00359869s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:48.980184689 +0000 UTC m=+149.800625303" watchObservedRunningTime="2025-12-27 07:44:49.00359869 +0000 UTC m=+149.824039284" Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.031845 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:49 crc kubenswrapper[4934]: E1227 07:44:49.032238 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:49.532223994 +0000 UTC m=+150.352664578 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.047290 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-w4pmg" event={"ID":"bd2142a6-e9cb-432d-adf9-61b5d3f3c5b1","Type":"ContainerStarted","Data":"caaaf68d7296c7d4f8eaeacaa6f1ddccdc549262f02122867f7ec9accf0bbcc3"} Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.081547 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-7tf92"] Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.111171 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj"] Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.144037 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:49 crc kubenswrapper[4934]: E1227 07:44:49.144398 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:49.644387192 +0000 UTC m=+150.464827786 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.193564 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22jvt"] Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.194155 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-5s9zf"] Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.253385 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:49 crc kubenswrapper[4934]: E1227 07:44:49.253647 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:49.753633764 +0000 UTC m=+150.574074358 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.354471 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:49 crc kubenswrapper[4934]: E1227 07:44:49.355135 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:49.855123467 +0000 UTC m=+150.675564061 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.369682 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9s2ls" podStartSLOduration=131.369667036 podStartE2EDuration="2m11.369667036s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:49.369318886 +0000 UTC m=+150.189759480" watchObservedRunningTime="2025-12-27 07:44:49.369667036 +0000 UTC m=+150.190107630" Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.371847 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" podStartSLOduration=131.37184084 podStartE2EDuration="2m11.37184084s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:49.231004557 +0000 UTC m=+150.051445151" watchObservedRunningTime="2025-12-27 07:44:49.37184084 +0000 UTC m=+150.192281444" Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.384039 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48"] Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.398530 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-md2nq"] Dec 27 07:44:49 crc kubenswrapper[4934]: W1227 07:44:49.447934 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0c610b9_d890_49ee_b321_a3f1efba4b0a.slice/crio-b72383b0fbd3f05fcc7e959a359c5681bd2499c6344ec067954dec23769b33ff WatchSource:0}: Error finding container b72383b0fbd3f05fcc7e959a359c5681bd2499c6344ec067954dec23769b33ff: Status 404 returned error can't find the container with id b72383b0fbd3f05fcc7e959a359c5681bd2499c6344ec067954dec23769b33ff Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.455879 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:49 crc kubenswrapper[4934]: E1227 07:44:49.456499 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:49.956484427 +0000 UTC m=+150.776925021 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.558546 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:49 crc kubenswrapper[4934]: E1227 07:44:49.558894 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:50.058883067 +0000 UTC m=+150.879323661 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.570658 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-8msj4" podStartSLOduration=131.570632213 podStartE2EDuration="2m11.570632213s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:49.557570548 +0000 UTC m=+150.378011142" watchObservedRunningTime="2025-12-27 07:44:49.570632213 +0000 UTC m=+150.391072807" Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.583959 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-spr46"] Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.661592 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-c2tqv"] Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.667065 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:49 crc kubenswrapper[4934]: E1227 07:44:49.667267 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:50.167247673 +0000 UTC m=+150.987688267 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.667577 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:49 crc kubenswrapper[4934]: E1227 07:44:49.668000 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:50.167988504 +0000 UTC m=+150.988429098 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.677869 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ph956" podStartSLOduration=131.677853505 podStartE2EDuration="2m11.677853505s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:49.676931228 +0000 UTC m=+150.497371822" watchObservedRunningTime="2025-12-27 07:44:49.677853505 +0000 UTC m=+150.498294099" Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.733586 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" podStartSLOduration=131.733572149 podStartE2EDuration="2m11.733572149s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:49.733472856 +0000 UTC m=+150.553913440" watchObservedRunningTime="2025-12-27 07:44:49.733572149 +0000 UTC m=+150.554012743" Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.747485 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xxrp6"] Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.797067 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:49 crc kubenswrapper[4934]: E1227 07:44:49.797406 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:50.297390371 +0000 UTC m=+151.117830965 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.820719 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-p8c2x" podStartSLOduration=131.820689908 podStartE2EDuration="2m11.820689908s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:49.761944835 +0000 UTC m=+150.582385429" watchObservedRunningTime="2025-12-27 07:44:49.820689908 +0000 UTC m=+150.641130522" Dec 27 07:44:49 crc kubenswrapper[4934]: I1227 07:44:49.907720 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:49 crc kubenswrapper[4934]: E1227 07:44:49.909030 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:50.409013923 +0000 UTC m=+151.229454517 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.008518 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:50 crc kubenswrapper[4934]: E1227 07:44:50.008885 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:50.508869058 +0000 UTC m=+151.329309652 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.110113 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:50 crc kubenswrapper[4934]: E1227 07:44:50.110448 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:50.610437224 +0000 UTC m=+151.430877818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.173603 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-5n6pv" event={"ID":"e06a3c18-5c45-4153-8b28-4cf918a55970","Type":"ContainerStarted","Data":"e0006608c5dcb6536dd1b7ed2cacaa0e2c659d0ed2497df7aa210061dad30d1b"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.173905 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-5n6pv" event={"ID":"e06a3c18-5c45-4153-8b28-4cf918a55970","Type":"ContainerStarted","Data":"407238c5a38786567391c25a113dba1101473bb5880c12abfeb3a83a368d6879"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.199678 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ndhjw" event={"ID":"b74aaf8a-71e4-4b9a-8212-c0f1cb2fd01b","Type":"ContainerStarted","Data":"52434fe6e2cf8d1bcef20097a9834593720033814d419af7aae9c0e0c08ebc72"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.208539 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-5n6pv" podStartSLOduration=6.208522176 podStartE2EDuration="6.208522176s" podCreationTimestamp="2025-12-27 07:44:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:50.207561098 +0000 UTC m=+151.028001692" watchObservedRunningTime="2025-12-27 07:44:50.208522176 +0000 UTC m=+151.028962780" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.210652 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:50 crc kubenswrapper[4934]: E1227 07:44:50.212704 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:50.712682849 +0000 UTC m=+151.533123443 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.237208 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-cpts5" event={"ID":"ac1de940-6952-466a-be96-52c43b1a90a2","Type":"ContainerStarted","Data":"8161ede993b45c1fb297f834b091010b2ab14be5128b47dc8516d0f92d8a73df"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.246235 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-c2tqv" event={"ID":"85b3562e-4570-4b9b-baa5-a20fe55f13ca","Type":"ContainerStarted","Data":"62252018b40b3591768f8a3aaf9f128cfeab9cb169f75f02acb959df4601070e"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.280115 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-cpts5" podStartSLOduration=132.280097127 podStartE2EDuration="2m12.280097127s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:50.279586832 +0000 UTC m=+151.100027426" watchObservedRunningTime="2025-12-27 07:44:50.280097127 +0000 UTC m=+151.100537721" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.311846 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:50 crc kubenswrapper[4934]: E1227 07:44:50.312157 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:50.812145902 +0000 UTC m=+151.632586496 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.320379 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-whhqx" event={"ID":"46527d26-eb38-494f-85f7-f5f56b8dfada","Type":"ContainerStarted","Data":"8b98eec201fe6f1db3c52f1d50655b19848ef9fb894737e734433dbcba121040"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.340362 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" event={"ID":"941a07b0-8984-43f4-8fcd-6cdc8c88f14a","Type":"ContainerStarted","Data":"64d5fb0ea26ea5c1ec2c6a50e37de1adb84c0c821483d1514a3979e9bb93a389"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.347228 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-md2nq" event={"ID":"66ea3a98-dc66-4baa-be63-dc53b00d3290","Type":"ContainerStarted","Data":"5235502829c21df1a4325579f63e99dcc65d9e383074b5342e4b357f483fad88"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.407335 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5phzx"] Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.409120 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5phzx" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.456773 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.461128 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn55p" event={"ID":"b2a55db0-a3bd-413f-a52f-75b36654377c","Type":"ContainerStarted","Data":"22f4508bc2a8af460e13f7e57bba84ef2f75ffc6cf76ee5ad2c13ad3fd9f425c"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.462297 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5phzx"] Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.462996 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:50 crc kubenswrapper[4934]: E1227 07:44:50.463207 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:50.963181917 +0000 UTC m=+151.783622511 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.463352 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:50 crc kubenswrapper[4934]: E1227 07:44:50.463693 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:50.963677562 +0000 UTC m=+151.784118156 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.531222 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wvjnl" event={"ID":"cb832863-8ab8-459a-8d73-3c464b8caf34","Type":"ContainerStarted","Data":"b42b054c4e8c282423751181b8a3d96a4c565daf017ad53fdaeba743af726090"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.531263 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wvjnl" event={"ID":"cb832863-8ab8-459a-8d73-3c464b8caf34","Type":"ContainerStarted","Data":"e4f26c879cf9081ba5bfff5b28e551ce843ac0dc97fe8580fc34889191c1596c"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.538678 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn55p" podStartSLOduration=132.538663243 podStartE2EDuration="2m12.538663243s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:50.534113109 +0000 UTC m=+151.354553713" watchObservedRunningTime="2025-12-27 07:44:50.538663243 +0000 UTC m=+151.359103837" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.572296 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.572540 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ks49\" (UniqueName: \"kubernetes.io/projected/ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20-kube-api-access-8ks49\") pod \"certified-operators-5phzx\" (UID: \"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20\") " pod="openshift-marketplace/certified-operators-5phzx" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.572612 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20-catalog-content\") pod \"certified-operators-5phzx\" (UID: \"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20\") " pod="openshift-marketplace/certified-operators-5phzx" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.572663 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20-utilities\") pod \"certified-operators-5phzx\" (UID: \"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20\") " pod="openshift-marketplace/certified-operators-5phzx" Dec 27 07:44:50 crc kubenswrapper[4934]: E1227 07:44:50.572777 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:51.072760339 +0000 UTC m=+151.893200933 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.623907 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-wvjnl" podStartSLOduration=132.623878885 podStartE2EDuration="2m12.623878885s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:50.615962902 +0000 UTC m=+151.436403496" watchObservedRunningTime="2025-12-27 07:44:50.623878885 +0000 UTC m=+151.444319469" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.654949 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hh662" event={"ID":"0b1dac22-8539-4f61-ac28-2f420c845a2a","Type":"ContainerStarted","Data":"1e3121d90dc5eecbb57325317b1208e68d7ee0b6c505942975189d868d307f4d"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.657353 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wtxbn"] Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.666526 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wtxbn" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.674857 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20-catalog-content\") pod \"certified-operators-5phzx\" (UID: \"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20\") " pod="openshift-marketplace/certified-operators-5phzx" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.674911 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20-utilities\") pod \"certified-operators-5phzx\" (UID: \"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20\") " pod="openshift-marketplace/certified-operators-5phzx" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.674974 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ks49\" (UniqueName: \"kubernetes.io/projected/ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20-kube-api-access-8ks49\") pod \"certified-operators-5phzx\" (UID: \"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20\") " pod="openshift-marketplace/certified-operators-5phzx" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.675018 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:50 crc kubenswrapper[4934]: E1227 07:44:50.675295 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:51.175284021 +0000 UTC m=+151.995724615 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.675692 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20-utilities\") pod \"certified-operators-5phzx\" (UID: \"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20\") " pod="openshift-marketplace/certified-operators-5phzx" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.676119 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.677023 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20-catalog-content\") pod \"certified-operators-5phzx\" (UID: \"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20\") " pod="openshift-marketplace/certified-operators-5phzx" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.679632 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-cpts5" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.692442 4934 patch_prober.go:28] interesting pod/router-default-5444994796-cpts5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 27 07:44:50 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Dec 27 07:44:50 crc kubenswrapper[4934]: [+]process-running ok Dec 27 07:44:50 crc kubenswrapper[4934]: healthz check failed Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.692486 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cpts5" podUID="ac1de940-6952-466a-be96-52c43b1a90a2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.702381 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22jvt" event={"ID":"9ef59357-6b1b-491e-a43d-3d436b09d4ff","Type":"ContainerStarted","Data":"c9e5e24ac6b31cf8b768c55d44b11db99ab214e3e7194df3f90c1a941a52baa9"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.712569 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wtxbn"] Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.723186 4934 generic.go:334] "Generic (PLEG): container finished" podID="4c71cabc-c13f-4ae8-a704-10efee323e35" containerID="acdd10c9b9f8c084769ae168b0255ad586ed3177ca5dd69a1ff48ebce83e22ac" exitCode=0 Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.723270 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" event={"ID":"4c71cabc-c13f-4ae8-a704-10efee323e35","Type":"ContainerDied","Data":"acdd10c9b9f8c084769ae168b0255ad586ed3177ca5dd69a1ff48ebce83e22ac"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.735009 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hh662" podStartSLOduration=132.734985022 podStartE2EDuration="2m12.734985022s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:50.719437404 +0000 UTC m=+151.539877998" watchObservedRunningTime="2025-12-27 07:44:50.734985022 +0000 UTC m=+151.555425616" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.740458 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ks49\" (UniqueName: \"kubernetes.io/projected/ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20-kube-api-access-8ks49\") pod \"certified-operators-5phzx\" (UID: \"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20\") " pod="openshift-marketplace/certified-operators-5phzx" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.752396 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zchqn" event={"ID":"e4f56101-c810-4e1f-96e8-c7cbcb2ec719","Type":"ContainerStarted","Data":"1e8e48b8c3b0e309fc773571d0b4517d63d3bf6394d169d104504687c58901ca"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.752430 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zchqn" event={"ID":"e4f56101-c810-4e1f-96e8-c7cbcb2ec719","Type":"ContainerStarted","Data":"f3ecd3107b0995a3ec688f9ba95d451bf1592a20cdefff55be6a8ae812b6f8b1"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.778837 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.779070 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b06ee67d-88e4-4c2f-b0a1-efdc7517c40d-utilities\") pod \"community-operators-wtxbn\" (UID: \"b06ee67d-88e4-4c2f-b0a1-efdc7517c40d\") " pod="openshift-marketplace/community-operators-wtxbn" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.779129 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xxg7\" (UniqueName: \"kubernetes.io/projected/b06ee67d-88e4-4c2f-b0a1-efdc7517c40d-kube-api-access-6xxg7\") pod \"community-operators-wtxbn\" (UID: \"b06ee67d-88e4-4c2f-b0a1-efdc7517c40d\") " pod="openshift-marketplace/community-operators-wtxbn" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.779155 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b06ee67d-88e4-4c2f-b0a1-efdc7517c40d-catalog-content\") pod \"community-operators-wtxbn\" (UID: \"b06ee67d-88e4-4c2f-b0a1-efdc7517c40d\") " pod="openshift-marketplace/community-operators-wtxbn" Dec 27 07:44:50 crc kubenswrapper[4934]: E1227 07:44:50.780095 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:51.280066482 +0000 UTC m=+152.100507076 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.786824 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22jvt" podStartSLOduration=132.7868057 podStartE2EDuration="2m12.7868057s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:50.785851092 +0000 UTC m=+151.606291686" watchObservedRunningTime="2025-12-27 07:44:50.7868057 +0000 UTC m=+151.607246294" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.800227 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5phzx" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.809151 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" event={"ID":"873d8371-2f45-409a-b18d-f4c7dbdabeea","Type":"ContainerStarted","Data":"8a321970113515818b71a40f56b2b6a4245c28879e8231a89ab1b5e79b9ad538"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.810455 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.824326 4934 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-t6q48 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" start-of-body= Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.824371 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" podUID="873d8371-2f45-409a-b18d-f4c7dbdabeea" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.838656 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2zfd9" event={"ID":"585041c4-cd72-4116-b7cf-5801881f349c","Type":"ContainerStarted","Data":"15ed79031ddd4d558dad16bf21ee465bf7483be6a6c702daff1c84355b877dc4"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.857295 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hrm6w" event={"ID":"cee9f5f3-5282-4ec5-94a1-fd193ea66f2e","Type":"ContainerStarted","Data":"18cb52aa120460892ceac5643348e3ed0302744066bb3ca8dea46c8c2bbb580b"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.864352 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gk922"] Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.865319 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gk922" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.880335 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b06ee67d-88e4-4c2f-b0a1-efdc7517c40d-utilities\") pod \"community-operators-wtxbn\" (UID: \"b06ee67d-88e4-4c2f-b0a1-efdc7517c40d\") " pod="openshift-marketplace/community-operators-wtxbn" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.880399 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xxg7\" (UniqueName: \"kubernetes.io/projected/b06ee67d-88e4-4c2f-b0a1-efdc7517c40d-kube-api-access-6xxg7\") pod \"community-operators-wtxbn\" (UID: \"b06ee67d-88e4-4c2f-b0a1-efdc7517c40d\") " pod="openshift-marketplace/community-operators-wtxbn" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.880418 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b06ee67d-88e4-4c2f-b0a1-efdc7517c40d-catalog-content\") pod \"community-operators-wtxbn\" (UID: \"b06ee67d-88e4-4c2f-b0a1-efdc7517c40d\") " pod="openshift-marketplace/community-operators-wtxbn" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.880453 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.881641 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" event={"ID":"4ea539b6-ffe0-4393-af8d-62579266697c","Type":"ContainerStarted","Data":"d2beced8cabeaaa6b7e2703afeac985c57d9358880695b3f97dfd7e847227b58"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.881670 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" event={"ID":"4ea539b6-ffe0-4393-af8d-62579266697c","Type":"ContainerStarted","Data":"93e84d621a3558275d3fe2ea90a30d79b81ee689bfcb3830c92a67d26986718a"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.882513 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.882916 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b06ee67d-88e4-4c2f-b0a1-efdc7517c40d-utilities\") pod \"community-operators-wtxbn\" (UID: \"b06ee67d-88e4-4c2f-b0a1-efdc7517c40d\") " pod="openshift-marketplace/community-operators-wtxbn" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.883919 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b06ee67d-88e4-4c2f-b0a1-efdc7517c40d-catalog-content\") pod \"community-operators-wtxbn\" (UID: \"b06ee67d-88e4-4c2f-b0a1-efdc7517c40d\") " pod="openshift-marketplace/community-operators-wtxbn" Dec 27 07:44:50 crc kubenswrapper[4934]: E1227 07:44:50.883923 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:51.383908974 +0000 UTC m=+152.204349568 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.892299 4934 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-r6jld container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" start-of-body= Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.892358 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" podUID="4ea539b6-ffe0-4393-af8d-62579266697c" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.904520 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gk922"] Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.904719 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-7tf92" event={"ID":"9552ddff-a57b-4311-a6d5-552a8b8efe19","Type":"ContainerStarted","Data":"7ff22095f73e8d56d36db7ccb8ad519d27dcc36fdadf653de72870ca83f55eda"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.924380 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xxg7\" (UniqueName: \"kubernetes.io/projected/b06ee67d-88e4-4c2f-b0a1-efdc7517c40d-kube-api-access-6xxg7\") pod \"community-operators-wtxbn\" (UID: \"b06ee67d-88e4-4c2f-b0a1-efdc7517c40d\") " pod="openshift-marketplace/community-operators-wtxbn" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.948649 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj" event={"ID":"ce5a8998-bb68-4764-aa7d-a47ec23c4514","Type":"ContainerStarted","Data":"2391eab3fa2dd459e11dcc806bc76371447e10bd62f1d5509f1e10edd4fe92af"} Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.981509 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-w4pmg" event={"ID":"bd2142a6-e9cb-432d-adf9-61b5d3f3c5b1","Type":"ContainerStarted","Data":"fb509cefa2a4d539dda5a87cab3f285781b2a70be0bf277341a3428904b465c5"} Dec 27 07:44:50 crc kubenswrapper[4934]: E1227 07:44:50.981750 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:51.481732479 +0000 UTC m=+152.302173073 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.981676 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.982120 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f3aa628-fe89-4de2-ad21-34b7d18c9033-catalog-content\") pod \"certified-operators-gk922\" (UID: \"5f3aa628-fe89-4de2-ad21-34b7d18c9033\") " pod="openshift-marketplace/certified-operators-gk922" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.982148 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f3aa628-fe89-4de2-ad21-34b7d18c9033-utilities\") pod \"certified-operators-gk922\" (UID: \"5f3aa628-fe89-4de2-ad21-34b7d18c9033\") " pod="openshift-marketplace/certified-operators-gk922" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.982188 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:50 crc kubenswrapper[4934]: I1227 07:44:50.982205 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvtqw\" (UniqueName: \"kubernetes.io/projected/5f3aa628-fe89-4de2-ad21-34b7d18c9033-kube-api-access-vvtqw\") pod \"certified-operators-gk922\" (UID: \"5f3aa628-fe89-4de2-ad21-34b7d18c9033\") " pod="openshift-marketplace/certified-operators-gk922" Dec 27 07:44:50 crc kubenswrapper[4934]: E1227 07:44:50.986042 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:51.486027266 +0000 UTC m=+152.306467880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.014507 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wtxbn" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.036783 4934 generic.go:334] "Generic (PLEG): container finished" podID="d812eb7a-ce30-4759-a558-fff3738068b6" containerID="7e54e67c565455f3d71ba2dc0b9aa36f2554621e4ad4ba198c7b78aeffecefd2" exitCode=0 Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.036944 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" event={"ID":"d812eb7a-ce30-4759-a558-fff3738068b6","Type":"ContainerDied","Data":"7e54e67c565455f3d71ba2dc0b9aa36f2554621e4ad4ba198c7b78aeffecefd2"} Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.036984 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gtrfs"] Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.038256 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gtrfs" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.079840 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gtrfs"] Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.082978 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.083210 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f3aa628-fe89-4de2-ad21-34b7d18c9033-utilities\") pod \"certified-operators-gk922\" (UID: \"5f3aa628-fe89-4de2-ad21-34b7d18c9033\") " pod="openshift-marketplace/certified-operators-gk922" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.083280 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvtqw\" (UniqueName: \"kubernetes.io/projected/5f3aa628-fe89-4de2-ad21-34b7d18c9033-kube-api-access-vvtqw\") pod \"certified-operators-gk922\" (UID: \"5f3aa628-fe89-4de2-ad21-34b7d18c9033\") " pod="openshift-marketplace/certified-operators-gk922" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.083462 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6544m\" (UniqueName: \"kubernetes.io/projected/c26588b0-e09e-4bac-bfd4-22289100341e-kube-api-access-6544m\") pod \"community-operators-gtrfs\" (UID: \"c26588b0-e09e-4bac-bfd4-22289100341e\") " pod="openshift-marketplace/community-operators-gtrfs" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.083507 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c26588b0-e09e-4bac-bfd4-22289100341e-catalog-content\") pod \"community-operators-gtrfs\" (UID: \"c26588b0-e09e-4bac-bfd4-22289100341e\") " pod="openshift-marketplace/community-operators-gtrfs" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.083575 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f3aa628-fe89-4de2-ad21-34b7d18c9033-catalog-content\") pod \"certified-operators-gk922\" (UID: \"5f3aa628-fe89-4de2-ad21-34b7d18c9033\") " pod="openshift-marketplace/certified-operators-gk922" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.083602 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c26588b0-e09e-4bac-bfd4-22289100341e-utilities\") pod \"community-operators-gtrfs\" (UID: \"c26588b0-e09e-4bac-bfd4-22289100341e\") " pod="openshift-marketplace/community-operators-gtrfs" Dec 27 07:44:51 crc kubenswrapper[4934]: E1227 07:44:51.084123 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:51.584102168 +0000 UTC m=+152.404542762 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.084707 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f3aa628-fe89-4de2-ad21-34b7d18c9033-utilities\") pod \"certified-operators-gk922\" (UID: \"5f3aa628-fe89-4de2-ad21-34b7d18c9033\") " pod="openshift-marketplace/certified-operators-gk922" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.088694 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f3aa628-fe89-4de2-ad21-34b7d18c9033-catalog-content\") pod \"certified-operators-gk922\" (UID: \"5f3aa628-fe89-4de2-ad21-34b7d18c9033\") " pod="openshift-marketplace/certified-operators-gk922" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.089768 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" event={"ID":"a0c610b9-d890-49ee-b321-a3f1efba4b0a","Type":"ContainerStarted","Data":"b72383b0fbd3f05fcc7e959a359c5681bd2499c6344ec067954dec23769b33ff"} Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.151294 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bmfr" event={"ID":"3da2af1b-ff3c-420c-af84-633eaddc9368","Type":"ContainerStarted","Data":"833797bda7e365802e41e973633717d837d091e840cb105708bf181769f32edf"} Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.151346 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bmfr" event={"ID":"3da2af1b-ff3c-420c-af84-633eaddc9368","Type":"ContainerStarted","Data":"38a6fa604405a4e9d14e9ba5fa20fd127fdff4100d24fe60661543bcd8fe4d66"} Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.155445 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvtqw\" (UniqueName: \"kubernetes.io/projected/5f3aa628-fe89-4de2-ad21-34b7d18c9033-kube-api-access-vvtqw\") pod \"certified-operators-gk922\" (UID: \"5f3aa628-fe89-4de2-ad21-34b7d18c9033\") " pod="openshift-marketplace/certified-operators-gk922" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.191351 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6544m\" (UniqueName: \"kubernetes.io/projected/c26588b0-e09e-4bac-bfd4-22289100341e-kube-api-access-6544m\") pod \"community-operators-gtrfs\" (UID: \"c26588b0-e09e-4bac-bfd4-22289100341e\") " pod="openshift-marketplace/community-operators-gtrfs" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.191407 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c26588b0-e09e-4bac-bfd4-22289100341e-catalog-content\") pod \"community-operators-gtrfs\" (UID: \"c26588b0-e09e-4bac-bfd4-22289100341e\") " pod="openshift-marketplace/community-operators-gtrfs" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.191453 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c26588b0-e09e-4bac-bfd4-22289100341e-utilities\") pod \"community-operators-gtrfs\" (UID: \"c26588b0-e09e-4bac-bfd4-22289100341e\") " pod="openshift-marketplace/community-operators-gtrfs" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.191489 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:51 crc kubenswrapper[4934]: E1227 07:44:51.191880 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:51.691862567 +0000 UTC m=+152.512303161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.192726 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-z76rm" event={"ID":"12f53d40-4cdb-41e5-b620-c85db6cb410a","Type":"ContainerStarted","Data":"ff8b55c3f61dffb232bedccd6b0c2a1d649d417639d0b3a067046a937418d093"} Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.198138 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c26588b0-e09e-4bac-bfd4-22289100341e-utilities\") pod \"community-operators-gtrfs\" (UID: \"c26588b0-e09e-4bac-bfd4-22289100341e\") " pod="openshift-marketplace/community-operators-gtrfs" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.209396 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c26588b0-e09e-4bac-bfd4-22289100341e-catalog-content\") pod \"community-operators-gtrfs\" (UID: \"c26588b0-e09e-4bac-bfd4-22289100341e\") " pod="openshift-marketplace/community-operators-gtrfs" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.252163 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" podStartSLOduration=133.235487703 podStartE2EDuration="2m13.235487703s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:51.225602622 +0000 UTC m=+152.046043226" watchObservedRunningTime="2025-12-27 07:44:51.235487703 +0000 UTC m=+152.055928297" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.265198 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hrm6w" podStartSLOduration=133.265158778 podStartE2EDuration="2m13.265158778s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:51.150986771 +0000 UTC m=+151.971427365" watchObservedRunningTime="2025-12-27 07:44:51.265158778 +0000 UTC m=+152.085599392" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.270436 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-jckgt" event={"ID":"753c599b-a401-481e-a5b3-e2dd6ed2c4e7","Type":"ContainerStarted","Data":"266ca8fe5b2103126717fbf60060271bbdf278ba0fdb38735bae74d5e0d42317"} Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.271158 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-jckgt" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.293724 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:51 crc kubenswrapper[4934]: E1227 07:44:51.301328 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:51.801290454 +0000 UTC m=+152.621731048 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.301991 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6544m\" (UniqueName: \"kubernetes.io/projected/c26588b0-e09e-4bac-bfd4-22289100341e-kube-api-access-6544m\") pod \"community-operators-gtrfs\" (UID: \"c26588b0-e09e-4bac-bfd4-22289100341e\") " pod="openshift-marketplace/community-operators-gtrfs" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.309260 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gk922" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.352381 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gdpgt" event={"ID":"d3ba307a-a762-4e65-aee8-d30964fadea9","Type":"ContainerStarted","Data":"7d22098eeac189e86f5f6a346723d09b0421a309ce88593019ec8ecbfa6d1293"} Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.353119 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-jckgt" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.358707 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zchqn" podStartSLOduration=133.342072857 podStartE2EDuration="2m13.342072857s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:51.293398661 +0000 UTC m=+152.113839255" watchObservedRunningTime="2025-12-27 07:44:51.342072857 +0000 UTC m=+152.162513451" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.393416 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gtrfs" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.399539 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" podStartSLOduration=133.399529341 podStartE2EDuration="2m13.399529341s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:51.359539172 +0000 UTC m=+152.179979766" watchObservedRunningTime="2025-12-27 07:44:51.399529341 +0000 UTC m=+152.219969935" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.406385 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:51 crc kubenswrapper[4934]: E1227 07:44:51.408031 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:51.908017172 +0000 UTC m=+152.728457766 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.414745 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zcl7d" event={"ID":"a1e4154b-2230-4499-b645-4903ea9593f1","Type":"ContainerStarted","Data":"51e91417e9fd5092bc736c627b996c6b0cd0f48c771e772240eb7fac21745a68"} Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.416633 4934 generic.go:334] "Generic (PLEG): container finished" podID="145ee02a-c7aa-4990-a288-e6f69f8ab030" containerID="449d76768df46be0c4dccc825a8edc2fe71c5cf51e03858194df1a62d62a1221" exitCode=0 Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.416669 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" event={"ID":"145ee02a-c7aa-4990-a288-e6f69f8ab030","Type":"ContainerDied","Data":"449d76768df46be0c4dccc825a8edc2fe71c5cf51e03858194df1a62d62a1221"} Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.431498 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-w4pmg" podStartSLOduration=132.431485824 podStartE2EDuration="2m12.431485824s" podCreationTimestamp="2025-12-27 07:42:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:51.431398111 +0000 UTC m=+152.251838715" watchObservedRunningTime="2025-12-27 07:44:51.431485824 +0000 UTC m=+152.251926408" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.449984 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ph956" event={"ID":"c16c1071-8350-4dda-b34b-183cad108e19","Type":"ContainerStarted","Data":"5b203d6dd2960455a976bc5e2579d889510b93181dd80c019b20b5bbae7ca4c0"} Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.472495 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-7tf92" podStartSLOduration=7.472467412 podStartE2EDuration="7.472467412s" podCreationTimestamp="2025-12-27 07:44:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:51.472360489 +0000 UTC m=+152.292801073" watchObservedRunningTime="2025-12-27 07:44:51.472467412 +0000 UTC m=+152.292908006" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.507561 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:51 crc kubenswrapper[4934]: E1227 07:44:51.513404 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:52.013381369 +0000 UTC m=+152.833821963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.515868 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:51 crc kubenswrapper[4934]: E1227 07:44:51.517645 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:52.017632664 +0000 UTC m=+152.838073258 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.545640 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gdpgt" podStartSLOduration=133.54562126 podStartE2EDuration="2m13.54562126s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:51.544523967 +0000 UTC m=+152.364964561" watchObservedRunningTime="2025-12-27 07:44:51.54562126 +0000 UTC m=+152.366061854" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.614343 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8jpzb" event={"ID":"ba604131-ba14-4754-8c5e-3e4fdccdfe64","Type":"ContainerStarted","Data":"d10fdb7d5ed35503b5f2aa3036031b5224a914a0805ed9d9dd9d04ccac4ee771"} Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.617532 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:51 crc kubenswrapper[4934]: E1227 07:44:51.617621 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:52.117606313 +0000 UTC m=+152.938046907 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.617708 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:51 crc kubenswrapper[4934]: E1227 07:44:51.617961 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:52.117952933 +0000 UTC m=+152.938393527 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.656681 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" event={"ID":"5548c900-e014-4c35-b62e-455cd82f5887","Type":"ContainerStarted","Data":"c59d13363d1636e34f915d67816d51af03491a24eef81601c054e63b8657f516"} Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.657520 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.659131 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-nr4vm" event={"ID":"5c9fc52e-279a-44a8-a779-17966e320a93","Type":"ContainerStarted","Data":"0aefca8133db189482edb880cfd8dc99f8d03f65b13d8732f033bb480b909160"} Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.696393 4934 patch_prober.go:28] interesting pod/router-default-5444994796-cpts5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 27 07:44:51 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Dec 27 07:44:51 crc kubenswrapper[4934]: [+]process-running ok Dec 27 07:44:51 crc kubenswrapper[4934]: healthz check failed Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.696448 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cpts5" podUID="ac1de940-6952-466a-be96-52c43b1a90a2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.697521 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-spr46" event={"ID":"2eec2888-f042-4de8-a6f7-b2bb2bb3db14","Type":"ContainerStarted","Data":"cc83f5c62f59c578bdff2a7e7ec28ed800c68b7d237566aa5d141477dcbfe469"} Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.697719 4934 patch_prober.go:28] interesting pod/downloads-7954f5f757-8msj4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.697746 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8msj4" podUID="a0d34185-bf08-456c-a5fc-3f0852d35a91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.698752 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.714470 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.714657 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.720202 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:51 crc kubenswrapper[4934]: E1227 07:44:51.720325 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:52.220303202 +0000 UTC m=+153.040743796 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.720785 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.731497 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" Dec 27 07:44:51 crc kubenswrapper[4934]: E1227 07:44:51.733395 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:52.233358697 +0000 UTC m=+153.053799291 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.823590 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-z76rm" podStartSLOduration=133.82356149700001 podStartE2EDuration="2m13.823561497s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:51.723343221 +0000 UTC m=+152.543783815" watchObservedRunningTime="2025-12-27 07:44:51.823561497 +0000 UTC m=+152.644002091" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.830690 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:51 crc kubenswrapper[4934]: E1227 07:44:51.832359 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:52.332342696 +0000 UTC m=+153.152783290 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.926384 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj" podStartSLOduration=133.926368859 podStartE2EDuration="2m13.926368859s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:51.88063369 +0000 UTC m=+152.701074284" watchObservedRunningTime="2025-12-27 07:44:51.926368859 +0000 UTC m=+152.746809453" Dec 27 07:44:51 crc kubenswrapper[4934]: I1227 07:44:51.934505 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:51 crc kubenswrapper[4934]: E1227 07:44:51.934801 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:52.434789867 +0000 UTC m=+153.255230461 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:51.991977 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6bmfr" podStartSLOduration=133.991961844 podStartE2EDuration="2m13.991961844s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:51.927242985 +0000 UTC m=+152.747683579" watchObservedRunningTime="2025-12-27 07:44:51.991961844 +0000 UTC m=+152.812402438" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:51.993454 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5phzx"] Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.008129 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-jckgt" podStartSLOduration=134.00811444 podStartE2EDuration="2m14.00811444s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:52.007516182 +0000 UTC m=+152.827956776" watchObservedRunningTime="2025-12-27 07:44:52.00811444 +0000 UTC m=+152.828555034" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.038130 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:52 crc kubenswrapper[4934]: E1227 07:44:52.038616 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:52.538600529 +0000 UTC m=+153.359041123 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.072192 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wtxbn"] Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.137858 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" podStartSLOduration=134.137844296 podStartE2EDuration="2m14.137844296s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:52.137447784 +0000 UTC m=+152.957888378" watchObservedRunningTime="2025-12-27 07:44:52.137844296 +0000 UTC m=+152.958284890" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.148196 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:52 crc kubenswrapper[4934]: E1227 07:44:52.148512 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:52.64849978 +0000 UTC m=+153.468940374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.254845 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:52 crc kubenswrapper[4934]: E1227 07:44:52.255198 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:52.755172946 +0000 UTC m=+153.575613540 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.255631 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:52 crc kubenswrapper[4934]: E1227 07:44:52.255953 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:52.755935579 +0000 UTC m=+153.576376173 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.295526 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-zcl7d" podStartSLOduration=134.295511226 podStartE2EDuration="2m14.295511226s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:52.293652231 +0000 UTC m=+153.114092825" watchObservedRunningTime="2025-12-27 07:44:52.295511226 +0000 UTC m=+153.115951820" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.358311 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:52 crc kubenswrapper[4934]: E1227 07:44:52.358560 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:52.858545125 +0000 UTC m=+153.678985719 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.399221 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8jpzb" podStartSLOduration=134.399209364 podStartE2EDuration="2m14.399209364s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:52.398457492 +0000 UTC m=+153.218898086" watchObservedRunningTime="2025-12-27 07:44:52.399209364 +0000 UTC m=+153.219649958" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.409382 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gtrfs"] Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.460610 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:52 crc kubenswrapper[4934]: E1227 07:44:52.460943 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:52.960932695 +0000 UTC m=+153.781373289 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.567125 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:52 crc kubenswrapper[4934]: E1227 07:44:52.567423 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:53.067408215 +0000 UTC m=+153.887848809 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.592073 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-spr46" podStartSLOduration=133.592057332 podStartE2EDuration="2m13.592057332s" podCreationTimestamp="2025-12-27 07:42:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:52.439286356 +0000 UTC m=+153.259726950" watchObservedRunningTime="2025-12-27 07:44:52.592057332 +0000 UTC m=+153.412497926" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.592647 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gk922"] Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.606067 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t4c2w"] Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.606954 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t4c2w" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.614354 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.622075 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t4c2w"] Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.671439 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.671821 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19b07e13-aa5d-4a0a-b6d9-254afc234025-utilities\") pod \"redhat-marketplace-t4c2w\" (UID: \"19b07e13-aa5d-4a0a-b6d9-254afc234025\") " pod="openshift-marketplace/redhat-marketplace-t4c2w" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.671853 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g8n6\" (UniqueName: \"kubernetes.io/projected/19b07e13-aa5d-4a0a-b6d9-254afc234025-kube-api-access-2g8n6\") pod \"redhat-marketplace-t4c2w\" (UID: \"19b07e13-aa5d-4a0a-b6d9-254afc234025\") " pod="openshift-marketplace/redhat-marketplace-t4c2w" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.671891 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19b07e13-aa5d-4a0a-b6d9-254afc234025-catalog-content\") pod \"redhat-marketplace-t4c2w\" (UID: \"19b07e13-aa5d-4a0a-b6d9-254afc234025\") " pod="openshift-marketplace/redhat-marketplace-t4c2w" Dec 27 07:44:52 crc kubenswrapper[4934]: E1227 07:44:52.672176 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:53.172164395 +0000 UTC m=+153.992604989 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.695531 4934 patch_prober.go:28] interesting pod/router-default-5444994796-cpts5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 27 07:44:52 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Dec 27 07:44:52 crc kubenswrapper[4934]: [+]process-running ok Dec 27 07:44:52 crc kubenswrapper[4934]: healthz check failed Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.695587 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cpts5" podUID="ac1de940-6952-466a-be96-52c43b1a90a2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.732659 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" event={"ID":"d812eb7a-ce30-4759-a558-fff3738068b6","Type":"ContainerStarted","Data":"e020e3e091320bf50fa8b691b84083173a057c645bc7c4b8364948b767ad37f3"} Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.756595 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wtxbn" event={"ID":"b06ee67d-88e4-4c2f-b0a1-efdc7517c40d","Type":"ContainerStarted","Data":"ea5f05850cd5912907fe94f65263850a595409b6306d5759b1b7d7cc5426fafc"} Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.763365 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" event={"ID":"145ee02a-c7aa-4990-a288-e6f69f8ab030","Type":"ContainerStarted","Data":"1f0f64f6af4087a97e599cf85a6ad6cf9797f98872e2f1da506adfd009fdf3ec"} Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.773588 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.773778 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19b07e13-aa5d-4a0a-b6d9-254afc234025-utilities\") pod \"redhat-marketplace-t4c2w\" (UID: \"19b07e13-aa5d-4a0a-b6d9-254afc234025\") " pod="openshift-marketplace/redhat-marketplace-t4c2w" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.773812 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g8n6\" (UniqueName: \"kubernetes.io/projected/19b07e13-aa5d-4a0a-b6d9-254afc234025-kube-api-access-2g8n6\") pod \"redhat-marketplace-t4c2w\" (UID: \"19b07e13-aa5d-4a0a-b6d9-254afc234025\") " pod="openshift-marketplace/redhat-marketplace-t4c2w" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.773852 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19b07e13-aa5d-4a0a-b6d9-254afc234025-catalog-content\") pod \"redhat-marketplace-t4c2w\" (UID: \"19b07e13-aa5d-4a0a-b6d9-254afc234025\") " pod="openshift-marketplace/redhat-marketplace-t4c2w" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.774243 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19b07e13-aa5d-4a0a-b6d9-254afc234025-catalog-content\") pod \"redhat-marketplace-t4c2w\" (UID: \"19b07e13-aa5d-4a0a-b6d9-254afc234025\") " pod="openshift-marketplace/redhat-marketplace-t4c2w" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.774308 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19b07e13-aa5d-4a0a-b6d9-254afc234025-utilities\") pod \"redhat-marketplace-t4c2w\" (UID: \"19b07e13-aa5d-4a0a-b6d9-254afc234025\") " pod="openshift-marketplace/redhat-marketplace-t4c2w" Dec 27 07:44:52 crc kubenswrapper[4934]: E1227 07:44:52.774392 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:53.274378669 +0000 UTC m=+154.094819263 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.838479 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-z76rm" event={"ID":"12f53d40-4cdb-41e5-b620-c85db6cb410a","Type":"ContainerStarted","Data":"62d02cbd25e41da050dec174b8ea86991ffca4a36b2ae246b6482152fa1bd055"} Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.843511 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g8n6\" (UniqueName: \"kubernetes.io/projected/19b07e13-aa5d-4a0a-b6d9-254afc234025-kube-api-access-2g8n6\") pod \"redhat-marketplace-t4c2w\" (UID: \"19b07e13-aa5d-4a0a-b6d9-254afc234025\") " pod="openshift-marketplace/redhat-marketplace-t4c2w" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.862137 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2zfd9" event={"ID":"585041c4-cd72-4116-b7cf-5801881f349c","Type":"ContainerStarted","Data":"7e4802e319f8b95db69fef935fcfdb158f940e0f41da587a73c2c94098f839bd"} Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.875105 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:52 crc kubenswrapper[4934]: E1227 07:44:52.875415 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:53.375401999 +0000 UTC m=+154.195842583 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.886793 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5phzx" event={"ID":"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20","Type":"ContainerStarted","Data":"12625e709f755da9b438242ba209f44a8633991928abc9dd8f53f07bdb02d4dd"} Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.886842 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5phzx" event={"ID":"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20","Type":"ContainerStarted","Data":"cff48171d21190f88ae52273299a0efe047ac067b630bb9c0c7b6af68d0a42b7"} Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.890317 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.895978 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" podStartSLOduration=133.895952575 podStartE2EDuration="2m13.895952575s" podCreationTimestamp="2025-12-27 07:42:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:52.827997461 +0000 UTC m=+153.648438065" watchObservedRunningTime="2025-12-27 07:44:52.895952575 +0000 UTC m=+153.716393159" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.896621 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-2zfd9" podStartSLOduration=134.896616884 podStartE2EDuration="2m14.896616884s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:52.895294725 +0000 UTC m=+153.715735329" watchObservedRunningTime="2025-12-27 07:44:52.896616884 +0000 UTC m=+153.717057478" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.905909 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-8jpzb" event={"ID":"ba604131-ba14-4754-8c5e-3e4fdccdfe64","Type":"ContainerStarted","Data":"6fbf8a005d0dd60aa977d1b458b95a95f50bc4e1d0fdf7e524db35365d5f89dd"} Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.925485 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" event={"ID":"941a07b0-8984-43f4-8fcd-6cdc8c88f14a","Type":"ContainerStarted","Data":"b80c8a2c584e6d6e80879c1af27f4192e3e3344b9730722a768ffc2832867f59"} Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.926520 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.929215 4934 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-xxrp6 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.929397 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" podUID="941a07b0-8984-43f4-8fcd-6cdc8c88f14a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.932449 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" event={"ID":"873d8371-2f45-409a-b18d-f4c7dbdabeea","Type":"ContainerStarted","Data":"4bf1d93325579a031c6fd1a9de6d9d6c91a118fcca02cf8d81a3e755107b73c5"} Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.950939 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.951598 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t4c2w" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.953951 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-c2tqv" event={"ID":"85b3562e-4570-4b9b-baa5-a20fe55f13ca","Type":"ContainerStarted","Data":"e23f32c43c8aa47e8d79d7f8632144b7dec5ca1bf53f050cc17bd6414a8e897e"} Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.954003 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-c2tqv" event={"ID":"85b3562e-4570-4b9b-baa5-a20fe55f13ca","Type":"ContainerStarted","Data":"2b66e3107f244f3385cbc85879bf460e0ff26e58347a773325b8c2749f69ae7a"} Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.954615 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-c2tqv" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.964288 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" podStartSLOduration=134.964262139 podStartE2EDuration="2m14.964262139s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:52.962667492 +0000 UTC m=+153.783108086" watchObservedRunningTime="2025-12-27 07:44:52.964262139 +0000 UTC m=+153.784702733" Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.975789 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:52 crc kubenswrapper[4934]: E1227 07:44:52.975905 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:53.475882272 +0000 UTC m=+154.296322876 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.976454 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:52 crc kubenswrapper[4934]: E1227 07:44:52.979140 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:53.479130568 +0000 UTC m=+154.299571162 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.988312 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ndhjw" event={"ID":"b74aaf8a-71e4-4b9a-8212-c0f1cb2fd01b","Type":"ContainerStarted","Data":"1ff3773c7ec6c7de0dc144e74de4a6cdc5f61da770bb531f678477853696a357"} Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.993988 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" event={"ID":"4c71cabc-c13f-4ae8-a704-10efee323e35","Type":"ContainerStarted","Data":"b631e608d82ae75f36a5a9fa3ea7f2bfcda667d3a9703769b16b2e70e44081be"} Dec 27 07:44:52 crc kubenswrapper[4934]: I1227 07:44:52.994234 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.017383 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-hrm6w" event={"ID":"cee9f5f3-5282-4ec5-94a1-fd193ea66f2e","Type":"ContainerStarted","Data":"80392aae99ba202b36388c8b826b4e2267e74853fddecc0f48311a1edb325ba8"} Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.027699 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtrfs" event={"ID":"c26588b0-e09e-4bac-bfd4-22289100341e","Type":"ContainerStarted","Data":"0728636b829e3e65d843ed46e682384caf901f66408a34a8522dd3473ac3c8ed"} Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.079328 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:53 crc kubenswrapper[4934]: E1227 07:44:53.081654 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:53.581629321 +0000 UTC m=+154.402069915 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.081700 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lzcth"] Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.093532 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lzcth" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.120105 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-whhqx" event={"ID":"46527d26-eb38-494f-85f7-f5f56b8dfada","Type":"ContainerStarted","Data":"ab14665cd63fae392ebcef592d1ad16e69bd20e4a2ba2ef2d11c62b7f0fa6cd2"} Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.179449 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-7tf92" event={"ID":"9552ddff-a57b-4311-a6d5-552a8b8efe19","Type":"ContainerStarted","Data":"23441b20ee3157a16ba092b4fce49dbbc80fd22878f0653f44fdd48b55c1d249"} Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.181824 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52a38d21-a579-4f37-be2f-856e2f837f83-utilities\") pod \"redhat-marketplace-lzcth\" (UID: \"52a38d21-a579-4f37-be2f-856e2f837f83\") " pod="openshift-marketplace/redhat-marketplace-lzcth" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.181875 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52a38d21-a579-4f37-be2f-856e2f837f83-catalog-content\") pod \"redhat-marketplace-lzcth\" (UID: \"52a38d21-a579-4f37-be2f-856e2f837f83\") " pod="openshift-marketplace/redhat-marketplace-lzcth" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.181905 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qwpc\" (UniqueName: \"kubernetes.io/projected/52a38d21-a579-4f37-be2f-856e2f837f83-kube-api-access-8qwpc\") pod \"redhat-marketplace-lzcth\" (UID: \"52a38d21-a579-4f37-be2f-856e2f837f83\") " pod="openshift-marketplace/redhat-marketplace-lzcth" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.181927 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:53 crc kubenswrapper[4934]: E1227 07:44:53.182228 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:53.682216007 +0000 UTC m=+154.502656601 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.191200 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lzcth"] Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.225864 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-c2tqv" podStartSLOduration=135.225821173 podStartE2EDuration="2m15.225821173s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:53.154370956 +0000 UTC m=+153.974811550" watchObservedRunningTime="2025-12-27 07:44:53.225821173 +0000 UTC m=+154.046261777" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.241934 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-ndhjw" podStartSLOduration=135.241904038 podStartE2EDuration="2m15.241904038s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:53.204471494 +0000 UTC m=+154.024912088" watchObservedRunningTime="2025-12-27 07:44:53.241904038 +0000 UTC m=+154.062344632" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.245837 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-spr46" event={"ID":"2eec2888-f042-4de8-a6f7-b2bb2bb3db14","Type":"ContainerStarted","Data":"eab617c6a13959da97d5fa29b54689ac69efd225e173ad278bc93118b4f8c772"} Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.268443 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" event={"ID":"a0c610b9-d890-49ee-b321-a3f1efba4b0a","Type":"ContainerStarted","Data":"64b2c7a5685c08b5a14c4b86aa2f6e5a2142cb223d7d4c4ff991df566ccc9ec2"} Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.283600 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.283833 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52a38d21-a579-4f37-be2f-856e2f837f83-utilities\") pod \"redhat-marketplace-lzcth\" (UID: \"52a38d21-a579-4f37-be2f-856e2f837f83\") " pod="openshift-marketplace/redhat-marketplace-lzcth" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.283860 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52a38d21-a579-4f37-be2f-856e2f837f83-catalog-content\") pod \"redhat-marketplace-lzcth\" (UID: \"52a38d21-a579-4f37-be2f-856e2f837f83\") " pod="openshift-marketplace/redhat-marketplace-lzcth" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.283898 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qwpc\" (UniqueName: \"kubernetes.io/projected/52a38d21-a579-4f37-be2f-856e2f837f83-kube-api-access-8qwpc\") pod \"redhat-marketplace-lzcth\" (UID: \"52a38d21-a579-4f37-be2f-856e2f837f83\") " pod="openshift-marketplace/redhat-marketplace-lzcth" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.284973 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52a38d21-a579-4f37-be2f-856e2f837f83-catalog-content\") pod \"redhat-marketplace-lzcth\" (UID: \"52a38d21-a579-4f37-be2f-856e2f837f83\") " pod="openshift-marketplace/redhat-marketplace-lzcth" Dec 27 07:44:53 crc kubenswrapper[4934]: E1227 07:44:53.285064 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:53.78504835 +0000 UTC m=+154.605488944 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.285329 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52a38d21-a579-4f37-be2f-856e2f837f83-utilities\") pod \"redhat-marketplace-lzcth\" (UID: \"52a38d21-a579-4f37-be2f-856e2f837f83\") " pod="openshift-marketplace/redhat-marketplace-lzcth" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.300234 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-mn55p" event={"ID":"b2a55db0-a3bd-413f-a52f-75b36654377c","Type":"ContainerStarted","Data":"740d0e450c8e6277244c4752f90d5fea38588a01ceddbf1cb9decc713d2bb297"} Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.321586 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gk922" event={"ID":"5f3aa628-fe89-4de2-ad21-34b7d18c9033","Type":"ContainerStarted","Data":"3b80adbdf4ad62e2df8313f61e990f658f1c3539e536bee70b1ecf9ccd6d3580"} Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.334147 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qwpc\" (UniqueName: \"kubernetes.io/projected/52a38d21-a579-4f37-be2f-856e2f837f83-kube-api-access-8qwpc\") pod \"redhat-marketplace-lzcth\" (UID: \"52a38d21-a579-4f37-be2f-856e2f837f83\") " pod="openshift-marketplace/redhat-marketplace-lzcth" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.336064 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" podStartSLOduration=135.336054045 podStartE2EDuration="2m15.336054045s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:53.3264033 +0000 UTC m=+154.146843894" watchObservedRunningTime="2025-12-27 07:44:53.336054045 +0000 UTC m=+154.156494639" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.347860 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22jvt" event={"ID":"9ef59357-6b1b-491e-a43d-3d436b09d4ff","Type":"ContainerStarted","Data":"6822ba3b08fe5a9e8ef466a95e607a1e0bdc1982c1590ac0585841f8e015b675"} Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.353144 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj" event={"ID":"ce5a8998-bb68-4764-aa7d-a47ec23c4514","Type":"ContainerStarted","Data":"cce8332487d739cc4965e08f8bea71277301b3c77c45c57b3a9fb5cbd3d65e86"} Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.355347 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-md2nq" event={"ID":"66ea3a98-dc66-4baa-be63-dc53b00d3290","Type":"ContainerStarted","Data":"8ca86a32bb3878b758d684b47b3cd7e6247093642198615a97d6c975649eb220"} Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.355367 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-md2nq" event={"ID":"66ea3a98-dc66-4baa-be63-dc53b00d3290","Type":"ContainerStarted","Data":"8596f37653f72d18a26c337ca6286d93a9ff57625b313bf1bf1738e3100e2065"} Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.355379 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-md2nq" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.378271 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-whhqx" podStartSLOduration=135.378244469 podStartE2EDuration="2m15.378244469s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:53.37592416 +0000 UTC m=+154.196364754" watchObservedRunningTime="2025-12-27 07:44:53.378244469 +0000 UTC m=+154.198685063" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.391107 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:53 crc kubenswrapper[4934]: E1227 07:44:53.392653 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:53.892618983 +0000 UTC m=+154.713059577 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.407185 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.435157 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-md2nq" podStartSLOduration=9.435140197 podStartE2EDuration="9.435140197s" podCreationTimestamp="2025-12-27 07:44:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:53.434427846 +0000 UTC m=+154.254868440" watchObservedRunningTime="2025-12-27 07:44:53.435140197 +0000 UTC m=+154.255580791" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.467312 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lzcth" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.495311 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:53 crc kubenswrapper[4934]: E1227 07:44:53.495434 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:53.995408814 +0000 UTC m=+154.815849408 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.496192 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:53 crc kubenswrapper[4934]: E1227 07:44:53.519283 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:54.019269098 +0000 UTC m=+154.839709692 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.600463 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:53 crc kubenswrapper[4934]: E1227 07:44:53.600842 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:54.100827473 +0000 UTC m=+154.921268067 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.618155 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xwb5s"] Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.625592 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwb5s" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.629731 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.655378 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xwb5s"] Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.674605 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t4c2w"] Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.690861 4934 patch_prober.go:28] interesting pod/router-default-5444994796-cpts5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 27 07:44:53 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Dec 27 07:44:53 crc kubenswrapper[4934]: [+]process-running ok Dec 27 07:44:53 crc kubenswrapper[4934]: healthz check failed Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.690922 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cpts5" podUID="ac1de940-6952-466a-be96-52c43b1a90a2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.703835 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/deaa2d15-bcaf-4568-8847-7bccae9205e0-utilities\") pod \"redhat-operators-xwb5s\" (UID: \"deaa2d15-bcaf-4568-8847-7bccae9205e0\") " pod="openshift-marketplace/redhat-operators-xwb5s" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.703895 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/deaa2d15-bcaf-4568-8847-7bccae9205e0-catalog-content\") pod \"redhat-operators-xwb5s\" (UID: \"deaa2d15-bcaf-4568-8847-7bccae9205e0\") " pod="openshift-marketplace/redhat-operators-xwb5s" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.703944 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gltj\" (UniqueName: \"kubernetes.io/projected/deaa2d15-bcaf-4568-8847-7bccae9205e0-kube-api-access-5gltj\") pod \"redhat-operators-xwb5s\" (UID: \"deaa2d15-bcaf-4568-8847-7bccae9205e0\") " pod="openshift-marketplace/redhat-operators-xwb5s" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.703972 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:53 crc kubenswrapper[4934]: E1227 07:44:53.704239 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:54.204226883 +0000 UTC m=+155.024667477 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.804566 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.804831 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/deaa2d15-bcaf-4568-8847-7bccae9205e0-catalog-content\") pod \"redhat-operators-xwb5s\" (UID: \"deaa2d15-bcaf-4568-8847-7bccae9205e0\") " pod="openshift-marketplace/redhat-operators-xwb5s" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.804886 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gltj\" (UniqueName: \"kubernetes.io/projected/deaa2d15-bcaf-4568-8847-7bccae9205e0-kube-api-access-5gltj\") pod \"redhat-operators-xwb5s\" (UID: \"deaa2d15-bcaf-4568-8847-7bccae9205e0\") " pod="openshift-marketplace/redhat-operators-xwb5s" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.804926 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/deaa2d15-bcaf-4568-8847-7bccae9205e0-utilities\") pod \"redhat-operators-xwb5s\" (UID: \"deaa2d15-bcaf-4568-8847-7bccae9205e0\") " pod="openshift-marketplace/redhat-operators-xwb5s" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.805376 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/deaa2d15-bcaf-4568-8847-7bccae9205e0-utilities\") pod \"redhat-operators-xwb5s\" (UID: \"deaa2d15-bcaf-4568-8847-7bccae9205e0\") " pod="openshift-marketplace/redhat-operators-xwb5s" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.805629 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/deaa2d15-bcaf-4568-8847-7bccae9205e0-catalog-content\") pod \"redhat-operators-xwb5s\" (UID: \"deaa2d15-bcaf-4568-8847-7bccae9205e0\") " pod="openshift-marketplace/redhat-operators-xwb5s" Dec 27 07:44:53 crc kubenswrapper[4934]: E1227 07:44:53.805693 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:54.305679065 +0000 UTC m=+155.126119659 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.851983 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gltj\" (UniqueName: \"kubernetes.io/projected/deaa2d15-bcaf-4568-8847-7bccae9205e0-kube-api-access-5gltj\") pod \"redhat-operators-xwb5s\" (UID: \"deaa2d15-bcaf-4568-8847-7bccae9205e0\") " pod="openshift-marketplace/redhat-operators-xwb5s" Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.906293 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:53 crc kubenswrapper[4934]: E1227 07:44:53.906583 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:54.406570651 +0000 UTC m=+155.227011235 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:53 crc kubenswrapper[4934]: I1227 07:44:53.948515 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwb5s" Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.002225 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lf8vn"] Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.007384 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lf8vn" Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.008686 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:54 crc kubenswrapper[4934]: E1227 07:44:54.009192 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:54.509175687 +0000 UTC m=+155.329616281 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.035977 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lf8vn"] Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.113688 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdz7v\" (UniqueName: \"kubernetes.io/projected/4d8b26f6-cfe5-44d8-89e2-92fff18c71a7-kube-api-access-zdz7v\") pod \"redhat-operators-lf8vn\" (UID: \"4d8b26f6-cfe5-44d8-89e2-92fff18c71a7\") " pod="openshift-marketplace/redhat-operators-lf8vn" Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.113730 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d8b26f6-cfe5-44d8-89e2-92fff18c71a7-catalog-content\") pod \"redhat-operators-lf8vn\" (UID: \"4d8b26f6-cfe5-44d8-89e2-92fff18c71a7\") " pod="openshift-marketplace/redhat-operators-lf8vn" Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.113955 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.114054 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d8b26f6-cfe5-44d8-89e2-92fff18c71a7-utilities\") pod \"redhat-operators-lf8vn\" (UID: \"4d8b26f6-cfe5-44d8-89e2-92fff18c71a7\") " pod="openshift-marketplace/redhat-operators-lf8vn" Dec 27 07:44:54 crc kubenswrapper[4934]: E1227 07:44:54.114353 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:54.614337398 +0000 UTC m=+155.434777992 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.214873 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:54 crc kubenswrapper[4934]: E1227 07:44:54.215021 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:54.715003286 +0000 UTC m=+155.535443880 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.215047 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdz7v\" (UniqueName: \"kubernetes.io/projected/4d8b26f6-cfe5-44d8-89e2-92fff18c71a7-kube-api-access-zdz7v\") pod \"redhat-operators-lf8vn\" (UID: \"4d8b26f6-cfe5-44d8-89e2-92fff18c71a7\") " pod="openshift-marketplace/redhat-operators-lf8vn" Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.215071 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d8b26f6-cfe5-44d8-89e2-92fff18c71a7-catalog-content\") pod \"redhat-operators-lf8vn\" (UID: \"4d8b26f6-cfe5-44d8-89e2-92fff18c71a7\") " pod="openshift-marketplace/redhat-operators-lf8vn" Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.215155 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.215189 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d8b26f6-cfe5-44d8-89e2-92fff18c71a7-utilities\") pod \"redhat-operators-lf8vn\" (UID: \"4d8b26f6-cfe5-44d8-89e2-92fff18c71a7\") " pod="openshift-marketplace/redhat-operators-lf8vn" Dec 27 07:44:54 crc kubenswrapper[4934]: E1227 07:44:54.215539 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:54.715531032 +0000 UTC m=+155.535971626 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.215568 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d8b26f6-cfe5-44d8-89e2-92fff18c71a7-utilities\") pod \"redhat-operators-lf8vn\" (UID: \"4d8b26f6-cfe5-44d8-89e2-92fff18c71a7\") " pod="openshift-marketplace/redhat-operators-lf8vn" Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.215726 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d8b26f6-cfe5-44d8-89e2-92fff18c71a7-catalog-content\") pod \"redhat-operators-lf8vn\" (UID: \"4d8b26f6-cfe5-44d8-89e2-92fff18c71a7\") " pod="openshift-marketplace/redhat-operators-lf8vn" Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.249197 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdz7v\" (UniqueName: \"kubernetes.io/projected/4d8b26f6-cfe5-44d8-89e2-92fff18c71a7-kube-api-access-zdz7v\") pod \"redhat-operators-lf8vn\" (UID: \"4d8b26f6-cfe5-44d8-89e2-92fff18c71a7\") " pod="openshift-marketplace/redhat-operators-lf8vn" Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.271457 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lzcth"] Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.315838 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:54 crc kubenswrapper[4934]: E1227 07:44:54.316104 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:54.816066177 +0000 UTC m=+155.636506771 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.316248 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:54 crc kubenswrapper[4934]: E1227 07:44:54.316550 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:54.816542721 +0000 UTC m=+155.636983315 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.351446 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lf8vn" Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.396253 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lzcth" event={"ID":"52a38d21-a579-4f37-be2f-856e2f837f83","Type":"ContainerStarted","Data":"fdae001289885a13b9cf483490888ea5141f8a90a0fef10f67613acf9f247051"} Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.398654 4934 generic.go:334] "Generic (PLEG): container finished" podID="5f3aa628-fe89-4de2-ad21-34b7d18c9033" containerID="a2c2c105ffe56da3716b3298c27e762726c6656192045ae2ebd324cba35b7b36" exitCode=0 Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.398703 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gk922" event={"ID":"5f3aa628-fe89-4de2-ad21-34b7d18c9033","Type":"ContainerDied","Data":"a2c2c105ffe56da3716b3298c27e762726c6656192045ae2ebd324cba35b7b36"} Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.403655 4934 generic.go:334] "Generic (PLEG): container finished" podID="ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" containerID="12625e709f755da9b438242ba209f44a8633991928abc9dd8f53f07bdb02d4dd" exitCode=0 Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.403717 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5phzx" event={"ID":"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20","Type":"ContainerDied","Data":"12625e709f755da9b438242ba209f44a8633991928abc9dd8f53f07bdb02d4dd"} Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.415970 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" event={"ID":"d812eb7a-ce30-4759-a558-fff3738068b6","Type":"ContainerStarted","Data":"f1a6798d030c1fb924087b58510b018f911117e0fd0b5d2e600cb8ec79628a0c"} Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.416972 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:54 crc kubenswrapper[4934]: E1227 07:44:54.417105 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:54.917074656 +0000 UTC m=+155.737515240 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.417222 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:54 crc kubenswrapper[4934]: E1227 07:44:54.417566 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:54.91755037 +0000 UTC m=+155.737990964 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.417590 4934 generic.go:334] "Generic (PLEG): container finished" podID="19b07e13-aa5d-4a0a-b6d9-254afc234025" containerID="5e32b748d4ec0850d4d6b5aeda4dc212f058207e2b0366c8279afa05c4bd4267" exitCode=0 Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.417632 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t4c2w" event={"ID":"19b07e13-aa5d-4a0a-b6d9-254afc234025","Type":"ContainerDied","Data":"5e32b748d4ec0850d4d6b5aeda4dc212f058207e2b0366c8279afa05c4bd4267"} Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.417648 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t4c2w" event={"ID":"19b07e13-aa5d-4a0a-b6d9-254afc234025","Type":"ContainerStarted","Data":"8bc7bab559fc2df71fda584267f3d57401360aa3d9305d85d74e16d4345f8855"} Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.430211 4934 generic.go:334] "Generic (PLEG): container finished" podID="c26588b0-e09e-4bac-bfd4-22289100341e" containerID="f6733cebfbca029ad99eaa4727ce1408064affd12a8ccd6008d9422faad893cf" exitCode=0 Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.430272 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtrfs" event={"ID":"c26588b0-e09e-4bac-bfd4-22289100341e","Type":"ContainerDied","Data":"f6733cebfbca029ad99eaa4727ce1408064affd12a8ccd6008d9422faad893cf"} Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.430502 4934 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.436293 4934 generic.go:334] "Generic (PLEG): container finished" podID="b06ee67d-88e4-4c2f-b0a1-efdc7517c40d" containerID="f7c5cadc3c57d85e2b7791f15387690e5c72a31597f5b0e6470db058f77c0138" exitCode=0 Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.436629 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wtxbn" event={"ID":"b06ee67d-88e4-4c2f-b0a1-efdc7517c40d","Type":"ContainerDied","Data":"f7c5cadc3c57d85e2b7791f15387690e5c72a31597f5b0e6470db058f77c0138"} Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.472129 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" podStartSLOduration=136.472109909 podStartE2EDuration="2m16.472109909s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:54.445760952 +0000 UTC m=+155.266201546" watchObservedRunningTime="2025-12-27 07:44:54.472109909 +0000 UTC m=+155.292550503" Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.502045 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" event={"ID":"a0c610b9-d890-49ee-b321-a3f1efba4b0a","Type":"ContainerStarted","Data":"68a8743bf3927c9a01cb563c33e68f613e059fa6e7c21de260ebf8e0df99111c"} Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.504055 4934 generic.go:334] "Generic (PLEG): container finished" podID="ce5a8998-bb68-4764-aa7d-a47ec23c4514" containerID="cce8332487d739cc4965e08f8bea71277301b3c77c45c57b3a9fb5cbd3d65e86" exitCode=0 Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.504266 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj" event={"ID":"ce5a8998-bb68-4764-aa7d-a47ec23c4514","Type":"ContainerDied","Data":"cce8332487d739cc4965e08f8bea71277301b3c77c45c57b3a9fb5cbd3d65e86"} Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.515671 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.518123 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:54 crc kubenswrapper[4934]: E1227 07:44:54.518579 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:55.018552329 +0000 UTC m=+155.838992923 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.522575 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.525829 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xwb5s"] Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.629412 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:54 crc kubenswrapper[4934]: E1227 07:44:54.632516 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-27 07:44:55.132502119 +0000 UTC m=+155.952942713 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-ln99b" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.689771 4934 patch_prober.go:28] interesting pod/router-default-5444994796-cpts5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 27 07:44:54 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Dec 27 07:44:54 crc kubenswrapper[4934]: [+]process-running ok Dec 27 07:44:54 crc kubenswrapper[4934]: healthz check failed Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.689822 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cpts5" podUID="ac1de940-6952-466a-be96-52c43b1a90a2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.690210 4934 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-27T07:44:54.430517212Z","Handler":null,"Name":""} Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.729990 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:54 crc kubenswrapper[4934]: E1227 07:44:54.730627 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-27 07:44:55.230611653 +0000 UTC m=+156.051052247 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.735642 4934 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.735677 4934 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.750393 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lf8vn"] Dec 27 07:44:54 crc kubenswrapper[4934]: W1227 07:44:54.793977 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4d8b26f6_cfe5_44d8_89e2_92fff18c71a7.slice/crio-18251686229c251ae27fa8528e64f334beca0386ecb27f257873591315f10162 WatchSource:0}: Error finding container 18251686229c251ae27fa8528e64f334beca0386ecb27f257873591315f10162: Status 404 returned error can't find the container with id 18251686229c251ae27fa8528e64f334beca0386ecb27f257873591315f10162 Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.831950 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.834811 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.834836 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.871150 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-ln99b\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.918731 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.932841 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 27 07:44:54 crc kubenswrapper[4934]: I1227 07:44:54.941622 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.080027 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.081325 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.084485 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.085657 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.087868 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.140643 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fca757e4-3dec-4183-8fad-0dae82804b83-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fca757e4-3dec-4183-8fad-0dae82804b83\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.140755 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fca757e4-3dec-4183-8fad-0dae82804b83-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fca757e4-3dec-4183-8fad-0dae82804b83\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.198410 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ln99b"] Dec 27 07:44:55 crc kubenswrapper[4934]: W1227 07:44:55.238731 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf922398_eb58_4f23_bfab_cfe9b90908c6.slice/crio-8f3350f6706e6a065226c875d01bf4bb025274fd53256fadead747481e7e1321 WatchSource:0}: Error finding container 8f3350f6706e6a065226c875d01bf4bb025274fd53256fadead747481e7e1321: Status 404 returned error can't find the container with id 8f3350f6706e6a065226c875d01bf4bb025274fd53256fadead747481e7e1321 Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.241420 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fca757e4-3dec-4183-8fad-0dae82804b83-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fca757e4-3dec-4183-8fad-0dae82804b83\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.241478 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fca757e4-3dec-4183-8fad-0dae82804b83-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fca757e4-3dec-4183-8fad-0dae82804b83\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.241847 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fca757e4-3dec-4183-8fad-0dae82804b83-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"fca757e4-3dec-4183-8fad-0dae82804b83\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.260953 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fca757e4-3dec-4183-8fad-0dae82804b83-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"fca757e4-3dec-4183-8fad-0dae82804b83\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.397102 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.483501 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.526434 4934 generic.go:334] "Generic (PLEG): container finished" podID="deaa2d15-bcaf-4568-8847-7bccae9205e0" containerID="fc9f95dec89a8df860380847f53fcda9d618a9c96cd22fe3801f1b278fbe883b" exitCode=0 Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.526638 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwb5s" event={"ID":"deaa2d15-bcaf-4568-8847-7bccae9205e0","Type":"ContainerDied","Data":"fc9f95dec89a8df860380847f53fcda9d618a9c96cd22fe3801f1b278fbe883b"} Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.526676 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwb5s" event={"ID":"deaa2d15-bcaf-4568-8847-7bccae9205e0","Type":"ContainerStarted","Data":"bbc7a319c5e881ac46200c346aeeaea1828f2c079ecedb7ff9d0ac5a7a97bcd6"} Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.547591 4934 generic.go:334] "Generic (PLEG): container finished" podID="4d8b26f6-cfe5-44d8-89e2-92fff18c71a7" containerID="b38f9f3b63ec18284f1b6d6f2c12cf4da720f9cde02731edab61293749bce9ee" exitCode=0 Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.547773 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lf8vn" event={"ID":"4d8b26f6-cfe5-44d8-89e2-92fff18c71a7","Type":"ContainerDied","Data":"b38f9f3b63ec18284f1b6d6f2c12cf4da720f9cde02731edab61293749bce9ee"} Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.547819 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lf8vn" event={"ID":"4d8b26f6-cfe5-44d8-89e2-92fff18c71a7","Type":"ContainerStarted","Data":"18251686229c251ae27fa8528e64f334beca0386ecb27f257873591315f10162"} Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.556037 4934 generic.go:334] "Generic (PLEG): container finished" podID="52a38d21-a579-4f37-be2f-856e2f837f83" containerID="597a65e4026ce1b1420029a2105f00a66a748b077e735ba0e448e75d82d38409" exitCode=0 Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.556107 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lzcth" event={"ID":"52a38d21-a579-4f37-be2f-856e2f837f83","Type":"ContainerDied","Data":"597a65e4026ce1b1420029a2105f00a66a748b077e735ba0e448e75d82d38409"} Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.576606 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" event={"ID":"a0c610b9-d890-49ee-b321-a3f1efba4b0a","Type":"ContainerStarted","Data":"b291d89c25898420ab637a9aea0cc0d0171b8de6a6b169cf803ef693fd1b5be8"} Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.576654 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" event={"ID":"a0c610b9-d890-49ee-b321-a3f1efba4b0a","Type":"ContainerStarted","Data":"d375af80d43976157fff37b2f69235aba68cdd1e5e6e0cdf65eb9cb4ed8ca15b"} Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.595420 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" event={"ID":"df922398-eb58-4f23-bfab-cfe9b90908c6","Type":"ContainerStarted","Data":"285a826b409cb26ca54b0f129254e7718f41a5ebe25c94a25c1ced5ddebe509d"} Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.595467 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" event={"ID":"df922398-eb58-4f23-bfab-cfe9b90908c6","Type":"ContainerStarted","Data":"8f3350f6706e6a065226c875d01bf4bb025274fd53256fadead747481e7e1321"} Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.610190 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" podStartSLOduration=11.610169673 podStartE2EDuration="11.610169673s" podCreationTimestamp="2025-12-27 07:44:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:55.600773796 +0000 UTC m=+156.421214400" watchObservedRunningTime="2025-12-27 07:44:55.610169673 +0000 UTC m=+156.430610267" Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.623193 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" podStartSLOduration=137.623172747 podStartE2EDuration="2m17.623172747s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:44:55.621544849 +0000 UTC m=+156.441985443" watchObservedRunningTime="2025-12-27 07:44:55.623172747 +0000 UTC m=+156.443613341" Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.688860 4934 patch_prober.go:28] interesting pod/router-default-5444994796-cpts5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 27 07:44:55 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Dec 27 07:44:55 crc kubenswrapper[4934]: [+]process-running ok Dec 27 07:44:55 crc kubenswrapper[4934]: healthz check failed Dec 27 07:44:55 crc kubenswrapper[4934]: I1227 07:44:55.688914 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cpts5" podUID="ac1de940-6952-466a-be96-52c43b1a90a2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.082116 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.153019 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5v6hs\" (UniqueName: \"kubernetes.io/projected/ce5a8998-bb68-4764-aa7d-a47ec23c4514-kube-api-access-5v6hs\") pod \"ce5a8998-bb68-4764-aa7d-a47ec23c4514\" (UID: \"ce5a8998-bb68-4764-aa7d-a47ec23c4514\") " Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.153059 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce5a8998-bb68-4764-aa7d-a47ec23c4514-config-volume\") pod \"ce5a8998-bb68-4764-aa7d-a47ec23c4514\" (UID: \"ce5a8998-bb68-4764-aa7d-a47ec23c4514\") " Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.153139 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce5a8998-bb68-4764-aa7d-a47ec23c4514-secret-volume\") pod \"ce5a8998-bb68-4764-aa7d-a47ec23c4514\" (UID: \"ce5a8998-bb68-4764-aa7d-a47ec23c4514\") " Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.154611 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce5a8998-bb68-4764-aa7d-a47ec23c4514-config-volume" (OuterVolumeSpecName: "config-volume") pod "ce5a8998-bb68-4764-aa7d-a47ec23c4514" (UID: "ce5a8998-bb68-4764-aa7d-a47ec23c4514"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.161372 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce5a8998-bb68-4764-aa7d-a47ec23c4514-kube-api-access-5v6hs" (OuterVolumeSpecName: "kube-api-access-5v6hs") pod "ce5a8998-bb68-4764-aa7d-a47ec23c4514" (UID: "ce5a8998-bb68-4764-aa7d-a47ec23c4514"). InnerVolumeSpecName "kube-api-access-5v6hs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.183842 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce5a8998-bb68-4764-aa7d-a47ec23c4514-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ce5a8998-bb68-4764-aa7d-a47ec23c4514" (UID: "ce5a8998-bb68-4764-aa7d-a47ec23c4514"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.210987 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.254170 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5v6hs\" (UniqueName: \"kubernetes.io/projected/ce5a8998-bb68-4764-aa7d-a47ec23c4514-kube-api-access-5v6hs\") on node \"crc\" DevicePath \"\"" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.254214 4934 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ce5a8998-bb68-4764-aa7d-a47ec23c4514-config-volume\") on node \"crc\" DevicePath \"\"" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.254224 4934 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ce5a8998-bb68-4764-aa7d-a47ec23c4514-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.605005 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj" event={"ID":"ce5a8998-bb68-4764-aa7d-a47ec23c4514","Type":"ContainerDied","Data":"2391eab3fa2dd459e11dcc806bc76371447e10bd62f1d5509f1e10edd4fe92af"} Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.605062 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.605096 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2391eab3fa2dd459e11dcc806bc76371447e10bd62f1d5509f1e10edd4fe92af" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.606565 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"fca757e4-3dec-4183-8fad-0dae82804b83","Type":"ContainerStarted","Data":"f2fbccb6621ef813e4c221da1afb72f873a6a8ba82ccfab112143556fabfcc44"} Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.606727 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.691586 4934 patch_prober.go:28] interesting pod/router-default-5444994796-cpts5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 27 07:44:56 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Dec 27 07:44:56 crc kubenswrapper[4934]: [+]process-running ok Dec 27 07:44:56 crc kubenswrapper[4934]: healthz check failed Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.691633 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cpts5" podUID="ac1de940-6952-466a-be96-52c43b1a90a2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.745739 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.745798 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.747617 4934 patch_prober.go:28] interesting pod/console-f9d7485db-p8c2x container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.747669 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-p8c2x" podUID="091c02c2-ae22-44e8-bf6b-64b69d6d9d83" containerName="console" probeResult="failure" output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.791925 4934 patch_prober.go:28] interesting pod/downloads-7954f5f757-8msj4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.791988 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8msj4" podUID="a0d34185-bf08-456c-a5fc-3f0852d35a91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.791998 4934 patch_prober.go:28] interesting pod/downloads-7954f5f757-8msj4 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.792059 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-8msj4" podUID="a0d34185-bf08-456c-a5fc-3f0852d35a91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.935030 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.935291 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.943228 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.953102 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.953148 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:56 crc kubenswrapper[4934]: I1227 07:44:56.962014 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:57 crc kubenswrapper[4934]: I1227 07:44:57.633312 4934 generic.go:334] "Generic (PLEG): container finished" podID="fca757e4-3dec-4183-8fad-0dae82804b83" containerID="b341f309f8d230bba46d7e349bf68ca19aaeeb03a363183e6b1773662c128600" exitCode=0 Dec 27 07:44:57 crc kubenswrapper[4934]: I1227 07:44:57.633469 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"fca757e4-3dec-4183-8fad-0dae82804b83","Type":"ContainerDied","Data":"b341f309f8d230bba46d7e349bf68ca19aaeeb03a363183e6b1773662c128600"} Dec 27 07:44:57 crc kubenswrapper[4934]: I1227 07:44:57.639180 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" Dec 27 07:44:57 crc kubenswrapper[4934]: I1227 07:44:57.640357 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" Dec 27 07:44:57 crc kubenswrapper[4934]: I1227 07:44:57.688193 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-cpts5" Dec 27 07:44:57 crc kubenswrapper[4934]: I1227 07:44:57.706119 4934 patch_prober.go:28] interesting pod/router-default-5444994796-cpts5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 27 07:44:57 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Dec 27 07:44:57 crc kubenswrapper[4934]: [+]process-running ok Dec 27 07:44:57 crc kubenswrapper[4934]: healthz check failed Dec 27 07:44:57 crc kubenswrapper[4934]: I1227 07:44:57.706178 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cpts5" podUID="ac1de940-6952-466a-be96-52c43b1a90a2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 07:44:58 crc kubenswrapper[4934]: I1227 07:44:58.685070 4934 patch_prober.go:28] interesting pod/router-default-5444994796-cpts5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 27 07:44:58 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Dec 27 07:44:58 crc kubenswrapper[4934]: [+]process-running ok Dec 27 07:44:58 crc kubenswrapper[4934]: healthz check failed Dec 27 07:44:58 crc kubenswrapper[4934]: I1227 07:44:58.685168 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cpts5" podUID="ac1de940-6952-466a-be96-52c43b1a90a2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:58.982116 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.016973 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fca757e4-3dec-4183-8fad-0dae82804b83-kube-api-access\") pod \"fca757e4-3dec-4183-8fad-0dae82804b83\" (UID: \"fca757e4-3dec-4183-8fad-0dae82804b83\") " Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.017068 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fca757e4-3dec-4183-8fad-0dae82804b83-kubelet-dir\") pod \"fca757e4-3dec-4183-8fad-0dae82804b83\" (UID: \"fca757e4-3dec-4183-8fad-0dae82804b83\") " Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.017271 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fca757e4-3dec-4183-8fad-0dae82804b83-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "fca757e4-3dec-4183-8fad-0dae82804b83" (UID: "fca757e4-3dec-4183-8fad-0dae82804b83"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.027475 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fca757e4-3dec-4183-8fad-0dae82804b83-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "fca757e4-3dec-4183-8fad-0dae82804b83" (UID: "fca757e4-3dec-4183-8fad-0dae82804b83"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.039640 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 27 07:44:59 crc kubenswrapper[4934]: E1227 07:44:59.039839 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fca757e4-3dec-4183-8fad-0dae82804b83" containerName="pruner" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.039849 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="fca757e4-3dec-4183-8fad-0dae82804b83" containerName="pruner" Dec 27 07:44:59 crc kubenswrapper[4934]: E1227 07:44:59.039871 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce5a8998-bb68-4764-aa7d-a47ec23c4514" containerName="collect-profiles" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.039877 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce5a8998-bb68-4764-aa7d-a47ec23c4514" containerName="collect-profiles" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.040008 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce5a8998-bb68-4764-aa7d-a47ec23c4514" containerName="collect-profiles" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.040025 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="fca757e4-3dec-4183-8fad-0dae82804b83" containerName="pruner" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.041075 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.045598 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.045820 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.046537 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.118694 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/024651fc-c731-4a71-b597-c069797a79a0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"024651fc-c731-4a71-b597-c069797a79a0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.118758 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/024651fc-c731-4a71-b597-c069797a79a0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"024651fc-c731-4a71-b597-c069797a79a0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.118949 4934 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fca757e4-3dec-4183-8fad-0dae82804b83-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.118976 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fca757e4-3dec-4183-8fad-0dae82804b83-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.219927 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/024651fc-c731-4a71-b597-c069797a79a0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"024651fc-c731-4a71-b597-c069797a79a0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.220027 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/024651fc-c731-4a71-b597-c069797a79a0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"024651fc-c731-4a71-b597-c069797a79a0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.220187 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/024651fc-c731-4a71-b597-c069797a79a0-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"024651fc-c731-4a71-b597-c069797a79a0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.239590 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/024651fc-c731-4a71-b597-c069797a79a0-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"024651fc-c731-4a71-b597-c069797a79a0\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.372536 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.668367 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"fca757e4-3dec-4183-8fad-0dae82804b83","Type":"ContainerDied","Data":"f2fbccb6621ef813e4c221da1afb72f873a6a8ba82ccfab112143556fabfcc44"} Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.668712 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2fbccb6621ef813e4c221da1afb72f873a6a8ba82ccfab112143556fabfcc44" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.668411 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.679226 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.682002 4934 patch_prober.go:28] interesting pod/router-default-5444994796-cpts5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 27 07:44:59 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Dec 27 07:44:59 crc kubenswrapper[4934]: [+]process-running ok Dec 27 07:44:59 crc kubenswrapper[4934]: healthz check failed Dec 27 07:44:59 crc kubenswrapper[4934]: I1227 07:44:59.682047 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cpts5" podUID="ac1de940-6952-466a-be96-52c43b1a90a2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 07:44:59 crc kubenswrapper[4934]: W1227 07:44:59.703419 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod024651fc_c731_4a71_b597_c069797a79a0.slice/crio-3d6dc9cce99a9f94b437221397ec9a1c15e8ecbea40ac3e0204e302bde01a6c8 WatchSource:0}: Error finding container 3d6dc9cce99a9f94b437221397ec9a1c15e8ecbea40ac3e0204e302bde01a6c8: Status 404 returned error can't find the container with id 3d6dc9cce99a9f94b437221397ec9a1c15e8ecbea40ac3e0204e302bde01a6c8 Dec 27 07:45:00 crc kubenswrapper[4934]: I1227 07:45:00.042064 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-md2nq" Dec 27 07:45:00 crc kubenswrapper[4934]: I1227 07:45:00.133260 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs"] Dec 27 07:45:00 crc kubenswrapper[4934]: I1227 07:45:00.134204 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs" Dec 27 07:45:00 crc kubenswrapper[4934]: I1227 07:45:00.135719 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 27 07:45:00 crc kubenswrapper[4934]: I1227 07:45:00.136767 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 27 07:45:00 crc kubenswrapper[4934]: I1227 07:45:00.144945 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs"] Dec 27 07:45:00 crc kubenswrapper[4934]: I1227 07:45:00.246810 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/20da3be4-5db1-43ec-9b86-5ffae5145bf8-secret-volume\") pod \"collect-profiles-29447025-g88fs\" (UID: \"20da3be4-5db1-43ec-9b86-5ffae5145bf8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs" Dec 27 07:45:00 crc kubenswrapper[4934]: I1227 07:45:00.247327 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8skq\" (UniqueName: \"kubernetes.io/projected/20da3be4-5db1-43ec-9b86-5ffae5145bf8-kube-api-access-x8skq\") pod \"collect-profiles-29447025-g88fs\" (UID: \"20da3be4-5db1-43ec-9b86-5ffae5145bf8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs" Dec 27 07:45:00 crc kubenswrapper[4934]: I1227 07:45:00.247349 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20da3be4-5db1-43ec-9b86-5ffae5145bf8-config-volume\") pod \"collect-profiles-29447025-g88fs\" (UID: \"20da3be4-5db1-43ec-9b86-5ffae5145bf8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs" Dec 27 07:45:00 crc kubenswrapper[4934]: I1227 07:45:00.349075 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8skq\" (UniqueName: \"kubernetes.io/projected/20da3be4-5db1-43ec-9b86-5ffae5145bf8-kube-api-access-x8skq\") pod \"collect-profiles-29447025-g88fs\" (UID: \"20da3be4-5db1-43ec-9b86-5ffae5145bf8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs" Dec 27 07:45:00 crc kubenswrapper[4934]: I1227 07:45:00.349127 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20da3be4-5db1-43ec-9b86-5ffae5145bf8-config-volume\") pod \"collect-profiles-29447025-g88fs\" (UID: \"20da3be4-5db1-43ec-9b86-5ffae5145bf8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs" Dec 27 07:45:00 crc kubenswrapper[4934]: I1227 07:45:00.349183 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/20da3be4-5db1-43ec-9b86-5ffae5145bf8-secret-volume\") pod \"collect-profiles-29447025-g88fs\" (UID: \"20da3be4-5db1-43ec-9b86-5ffae5145bf8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs" Dec 27 07:45:00 crc kubenswrapper[4934]: I1227 07:45:00.350506 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20da3be4-5db1-43ec-9b86-5ffae5145bf8-config-volume\") pod \"collect-profiles-29447025-g88fs\" (UID: \"20da3be4-5db1-43ec-9b86-5ffae5145bf8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs" Dec 27 07:45:00 crc kubenswrapper[4934]: I1227 07:45:00.386164 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8skq\" (UniqueName: \"kubernetes.io/projected/20da3be4-5db1-43ec-9b86-5ffae5145bf8-kube-api-access-x8skq\") pod \"collect-profiles-29447025-g88fs\" (UID: \"20da3be4-5db1-43ec-9b86-5ffae5145bf8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs" Dec 27 07:45:00 crc kubenswrapper[4934]: I1227 07:45:00.392416 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/20da3be4-5db1-43ec-9b86-5ffae5145bf8-secret-volume\") pod \"collect-profiles-29447025-g88fs\" (UID: \"20da3be4-5db1-43ec-9b86-5ffae5145bf8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs" Dec 27 07:45:00 crc kubenswrapper[4934]: I1227 07:45:00.455642 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs" Dec 27 07:45:00 crc kubenswrapper[4934]: I1227 07:45:00.683301 4934 patch_prober.go:28] interesting pod/router-default-5444994796-cpts5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 27 07:45:00 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Dec 27 07:45:00 crc kubenswrapper[4934]: [+]process-running ok Dec 27 07:45:00 crc kubenswrapper[4934]: healthz check failed Dec 27 07:45:00 crc kubenswrapper[4934]: I1227 07:45:00.683683 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cpts5" podUID="ac1de940-6952-466a-be96-52c43b1a90a2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 07:45:00 crc kubenswrapper[4934]: I1227 07:45:00.685894 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"024651fc-c731-4a71-b597-c069797a79a0","Type":"ContainerStarted","Data":"3d6dc9cce99a9f94b437221397ec9a1c15e8ecbea40ac3e0204e302bde01a6c8"} Dec 27 07:45:00 crc kubenswrapper[4934]: I1227 07:45:00.764608 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs\") pod \"network-metrics-daemon-jjlqg\" (UID: \"304709d3-05ab-4d27-8722-f94053159a29\") " pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:45:00 crc kubenswrapper[4934]: I1227 07:45:00.769125 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/304709d3-05ab-4d27-8722-f94053159a29-metrics-certs\") pod \"network-metrics-daemon-jjlqg\" (UID: \"304709d3-05ab-4d27-8722-f94053159a29\") " pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:45:01 crc kubenswrapper[4934]: I1227 07:45:01.010929 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs"] Dec 27 07:45:01 crc kubenswrapper[4934]: I1227 07:45:01.030522 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-jjlqg" Dec 27 07:45:01 crc kubenswrapper[4934]: I1227 07:45:01.387413 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-jjlqg"] Dec 27 07:45:01 crc kubenswrapper[4934]: I1227 07:45:01.681956 4934 patch_prober.go:28] interesting pod/router-default-5444994796-cpts5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 27 07:45:01 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Dec 27 07:45:01 crc kubenswrapper[4934]: [+]process-running ok Dec 27 07:45:01 crc kubenswrapper[4934]: healthz check failed Dec 27 07:45:01 crc kubenswrapper[4934]: I1227 07:45:01.682025 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cpts5" podUID="ac1de940-6952-466a-be96-52c43b1a90a2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 07:45:01 crc kubenswrapper[4934]: I1227 07:45:01.693136 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs" event={"ID":"20da3be4-5db1-43ec-9b86-5ffae5145bf8","Type":"ContainerStarted","Data":"79df9ef1359724aa621eb4d4adf6d48f6ebb9cfbe3ea82ab41ed4b032a1e6f76"} Dec 27 07:45:01 crc kubenswrapper[4934]: I1227 07:45:01.694678 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" event={"ID":"304709d3-05ab-4d27-8722-f94053159a29","Type":"ContainerStarted","Data":"e06b55bbd02121105f7219c37e778af73b6fe1fa29e99faa23917c67490587b3"} Dec 27 07:45:01 crc kubenswrapper[4934]: I1227 07:45:01.695767 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"024651fc-c731-4a71-b597-c069797a79a0","Type":"ContainerStarted","Data":"6d402ef57d266bc049558c01a3d0d656b678b7058562018873c9e7026ec85610"} Dec 27 07:45:01 crc kubenswrapper[4934]: I1227 07:45:01.712732 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.712716172 podStartE2EDuration="2.712716172s" podCreationTimestamp="2025-12-27 07:44:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:45:01.71060966 +0000 UTC m=+162.531050264" watchObservedRunningTime="2025-12-27 07:45:01.712716172 +0000 UTC m=+162.533156766" Dec 27 07:45:02 crc kubenswrapper[4934]: I1227 07:45:02.683069 4934 patch_prober.go:28] interesting pod/router-default-5444994796-cpts5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 27 07:45:02 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Dec 27 07:45:02 crc kubenswrapper[4934]: [+]process-running ok Dec 27 07:45:02 crc kubenswrapper[4934]: healthz check failed Dec 27 07:45:02 crc kubenswrapper[4934]: I1227 07:45:02.683171 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cpts5" podUID="ac1de940-6952-466a-be96-52c43b1a90a2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 07:45:02 crc kubenswrapper[4934]: I1227 07:45:02.704780 4934 generic.go:334] "Generic (PLEG): container finished" podID="024651fc-c731-4a71-b597-c069797a79a0" containerID="6d402ef57d266bc049558c01a3d0d656b678b7058562018873c9e7026ec85610" exitCode=0 Dec 27 07:45:02 crc kubenswrapper[4934]: I1227 07:45:02.705242 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"024651fc-c731-4a71-b597-c069797a79a0","Type":"ContainerDied","Data":"6d402ef57d266bc049558c01a3d0d656b678b7058562018873c9e7026ec85610"} Dec 27 07:45:02 crc kubenswrapper[4934]: I1227 07:45:02.707225 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs" event={"ID":"20da3be4-5db1-43ec-9b86-5ffae5145bf8","Type":"ContainerStarted","Data":"6852ddecc1458ae49f48e96d8d6f3450fe964be22e0c1a48aa7b5a131b953f19"} Dec 27 07:45:02 crc kubenswrapper[4934]: I1227 07:45:02.708622 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" event={"ID":"304709d3-05ab-4d27-8722-f94053159a29","Type":"ContainerStarted","Data":"08def858587d0394141bdad2b88f95d540315b8594210fff9008a6a9bdc440db"} Dec 27 07:45:03 crc kubenswrapper[4934]: I1227 07:45:03.682518 4934 patch_prober.go:28] interesting pod/router-default-5444994796-cpts5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 27 07:45:03 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Dec 27 07:45:03 crc kubenswrapper[4934]: [+]process-running ok Dec 27 07:45:03 crc kubenswrapper[4934]: healthz check failed Dec 27 07:45:03 crc kubenswrapper[4934]: I1227 07:45:03.682927 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cpts5" podUID="ac1de940-6952-466a-be96-52c43b1a90a2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 07:45:03 crc kubenswrapper[4934]: I1227 07:45:03.716154 4934 generic.go:334] "Generic (PLEG): container finished" podID="20da3be4-5db1-43ec-9b86-5ffae5145bf8" containerID="6852ddecc1458ae49f48e96d8d6f3450fe964be22e0c1a48aa7b5a131b953f19" exitCode=0 Dec 27 07:45:03 crc kubenswrapper[4934]: I1227 07:45:03.716626 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs" event={"ID":"20da3be4-5db1-43ec-9b86-5ffae5145bf8","Type":"ContainerDied","Data":"6852ddecc1458ae49f48e96d8d6f3450fe964be22e0c1a48aa7b5a131b953f19"} Dec 27 07:45:04 crc kubenswrapper[4934]: I1227 07:45:04.097409 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 27 07:45:04 crc kubenswrapper[4934]: I1227 07:45:04.209380 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/024651fc-c731-4a71-b597-c069797a79a0-kubelet-dir\") pod \"024651fc-c731-4a71-b597-c069797a79a0\" (UID: \"024651fc-c731-4a71-b597-c069797a79a0\") " Dec 27 07:45:04 crc kubenswrapper[4934]: I1227 07:45:04.209431 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/024651fc-c731-4a71-b597-c069797a79a0-kube-api-access\") pod \"024651fc-c731-4a71-b597-c069797a79a0\" (UID: \"024651fc-c731-4a71-b597-c069797a79a0\") " Dec 27 07:45:04 crc kubenswrapper[4934]: I1227 07:45:04.209701 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/024651fc-c731-4a71-b597-c069797a79a0-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "024651fc-c731-4a71-b597-c069797a79a0" (UID: "024651fc-c731-4a71-b597-c069797a79a0"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:45:04 crc kubenswrapper[4934]: I1227 07:45:04.219317 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/024651fc-c731-4a71-b597-c069797a79a0-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "024651fc-c731-4a71-b597-c069797a79a0" (UID: "024651fc-c731-4a71-b597-c069797a79a0"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:45:04 crc kubenswrapper[4934]: I1227 07:45:04.310819 4934 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/024651fc-c731-4a71-b597-c069797a79a0-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 27 07:45:04 crc kubenswrapper[4934]: I1227 07:45:04.310859 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/024651fc-c731-4a71-b597-c069797a79a0-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 27 07:45:04 crc kubenswrapper[4934]: I1227 07:45:04.681948 4934 patch_prober.go:28] interesting pod/router-default-5444994796-cpts5 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 27 07:45:04 crc kubenswrapper[4934]: [-]has-synced failed: reason withheld Dec 27 07:45:04 crc kubenswrapper[4934]: [+]process-running ok Dec 27 07:45:04 crc kubenswrapper[4934]: healthz check failed Dec 27 07:45:04 crc kubenswrapper[4934]: I1227 07:45:04.682014 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cpts5" podUID="ac1de940-6952-466a-be96-52c43b1a90a2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 07:45:04 crc kubenswrapper[4934]: I1227 07:45:04.725366 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-jjlqg" event={"ID":"304709d3-05ab-4d27-8722-f94053159a29","Type":"ContainerStarted","Data":"38bc1bde68aa5f87dd0d733ff8c57eef9719a0bccc9997f096d7e1ddeadba7f4"} Dec 27 07:45:04 crc kubenswrapper[4934]: I1227 07:45:04.729307 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 27 07:45:04 crc kubenswrapper[4934]: I1227 07:45:04.729730 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"024651fc-c731-4a71-b597-c069797a79a0","Type":"ContainerDied","Data":"3d6dc9cce99a9f94b437221397ec9a1c15e8ecbea40ac3e0204e302bde01a6c8"} Dec 27 07:45:04 crc kubenswrapper[4934]: I1227 07:45:04.729808 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d6dc9cce99a9f94b437221397ec9a1c15e8ecbea40ac3e0204e302bde01a6c8" Dec 27 07:45:04 crc kubenswrapper[4934]: I1227 07:45:04.744358 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-jjlqg" podStartSLOduration=146.744331213 podStartE2EDuration="2m26.744331213s" podCreationTimestamp="2025-12-27 07:42:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:45:04.7418597 +0000 UTC m=+165.562300294" watchObservedRunningTime="2025-12-27 07:45:04.744331213 +0000 UTC m=+165.564771807" Dec 27 07:45:05 crc kubenswrapper[4934]: I1227 07:45:05.683109 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-cpts5" Dec 27 07:45:05 crc kubenswrapper[4934]: I1227 07:45:05.686480 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-cpts5" Dec 27 07:45:06 crc kubenswrapper[4934]: I1227 07:45:06.745356 4934 patch_prober.go:28] interesting pod/console-f9d7485db-p8c2x container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Dec 27 07:45:06 crc kubenswrapper[4934]: I1227 07:45:06.745420 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-p8c2x" podUID="091c02c2-ae22-44e8-bf6b-64b69d6d9d83" containerName="console" probeResult="failure" output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" Dec 27 07:45:06 crc kubenswrapper[4934]: I1227 07:45:06.799159 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-8msj4" Dec 27 07:45:10 crc kubenswrapper[4934]: I1227 07:45:10.444748 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8npds"] Dec 27 07:45:10 crc kubenswrapper[4934]: I1227 07:45:10.445396 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" podUID="5548c900-e014-4c35-b62e-455cd82f5887" containerName="controller-manager" containerID="cri-o://c59d13363d1636e34f915d67816d51af03491a24eef81601c054e63b8657f516" gracePeriod=30 Dec 27 07:45:10 crc kubenswrapper[4934]: I1227 07:45:10.475108 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc"] Dec 27 07:45:10 crc kubenswrapper[4934]: I1227 07:45:10.475517 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" podUID="959b5ca2-d235-4457-98ec-fe077dfcdc6f" containerName="route-controller-manager" containerID="cri-o://b6f570194919b76e785ecb947ba89694a8df2a47ab393bdaaf2e4cf367a713c1" gracePeriod=30 Dec 27 07:45:12 crc kubenswrapper[4934]: I1227 07:45:12.813650 4934 generic.go:334] "Generic (PLEG): container finished" podID="959b5ca2-d235-4457-98ec-fe077dfcdc6f" containerID="b6f570194919b76e785ecb947ba89694a8df2a47ab393bdaaf2e4cf367a713c1" exitCode=0 Dec 27 07:45:12 crc kubenswrapper[4934]: I1227 07:45:12.813752 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" event={"ID":"959b5ca2-d235-4457-98ec-fe077dfcdc6f","Type":"ContainerDied","Data":"b6f570194919b76e785ecb947ba89694a8df2a47ab393bdaaf2e4cf367a713c1"} Dec 27 07:45:12 crc kubenswrapper[4934]: I1227 07:45:12.815866 4934 generic.go:334] "Generic (PLEG): container finished" podID="5548c900-e014-4c35-b62e-455cd82f5887" containerID="c59d13363d1636e34f915d67816d51af03491a24eef81601c054e63b8657f516" exitCode=0 Dec 27 07:45:12 crc kubenswrapper[4934]: I1227 07:45:12.815909 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" event={"ID":"5548c900-e014-4c35-b62e-455cd82f5887","Type":"ContainerDied","Data":"c59d13363d1636e34f915d67816d51af03491a24eef81601c054e63b8657f516"} Dec 27 07:45:14 crc kubenswrapper[4934]: I1227 07:45:14.927119 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:45:15 crc kubenswrapper[4934]: I1227 07:45:15.330626 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 07:45:15 crc kubenswrapper[4934]: I1227 07:45:15.330734 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 07:45:17 crc kubenswrapper[4934]: I1227 07:45:17.115863 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:45:17 crc kubenswrapper[4934]: I1227 07:45:17.120975 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:45:17 crc kubenswrapper[4934]: I1227 07:45:17.217119 4934 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-8npds container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Dec 27 07:45:17 crc kubenswrapper[4934]: I1227 07:45:17.217193 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" podUID="5548c900-e014-4c35-b62e-455cd82f5887" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" Dec 27 07:45:17 crc kubenswrapper[4934]: I1227 07:45:17.801464 4934 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-r5zxc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 07:45:17 crc kubenswrapper[4934]: I1227 07:45:17.801931 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" podUID="959b5ca2-d235-4457-98ec-fe077dfcdc6f" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 07:45:25 crc kubenswrapper[4934]: I1227 07:45:25.578325 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 27 07:45:27 crc kubenswrapper[4934]: I1227 07:45:27.217005 4934 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-8npds container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Dec 27 07:45:27 crc kubenswrapper[4934]: I1227 07:45:27.217240 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" podUID="5548c900-e014-4c35-b62e-455cd82f5887" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" Dec 27 07:45:27 crc kubenswrapper[4934]: I1227 07:45:27.801037 4934 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-r5zxc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 07:45:27 crc kubenswrapper[4934]: I1227 07:45:27.801158 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" podUID="959b5ca2-d235-4457-98ec-fe077dfcdc6f" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 07:45:27 crc kubenswrapper[4934]: I1227 07:45:27.934412 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-c2tqv" Dec 27 07:45:34 crc kubenswrapper[4934]: I1227 07:45:34.429077 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 27 07:45:34 crc kubenswrapper[4934]: E1227 07:45:34.429831 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="024651fc-c731-4a71-b597-c069797a79a0" containerName="pruner" Dec 27 07:45:34 crc kubenswrapper[4934]: I1227 07:45:34.429842 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="024651fc-c731-4a71-b597-c069797a79a0" containerName="pruner" Dec 27 07:45:34 crc kubenswrapper[4934]: I1227 07:45:34.429950 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="024651fc-c731-4a71-b597-c069797a79a0" containerName="pruner" Dec 27 07:45:34 crc kubenswrapper[4934]: I1227 07:45:34.430422 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 27 07:45:34 crc kubenswrapper[4934]: I1227 07:45:34.432654 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 27 07:45:34 crc kubenswrapper[4934]: I1227 07:45:34.432817 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 27 07:45:34 crc kubenswrapper[4934]: I1227 07:45:34.443678 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 27 07:45:34 crc kubenswrapper[4934]: I1227 07:45:34.604389 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 27 07:45:34 crc kubenswrapper[4934]: I1227 07:45:34.604474 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 27 07:45:34 crc kubenswrapper[4934]: I1227 07:45:34.706230 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 27 07:45:34 crc kubenswrapper[4934]: I1227 07:45:34.706286 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 27 07:45:34 crc kubenswrapper[4934]: I1227 07:45:34.706350 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 27 07:45:34 crc kubenswrapper[4934]: I1227 07:45:34.726062 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 27 07:45:34 crc kubenswrapper[4934]: I1227 07:45:34.836433 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.685170 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs" Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.690667 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" Dec 27 07:45:36 crc kubenswrapper[4934]: E1227 07:45:36.765642 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 27 07:45:36 crc kubenswrapper[4934]: E1227 07:45:36.769756 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6xxg7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-wtxbn_openshift-marketplace(b06ee67d-88e4-4c2f-b0a1-efdc7517c40d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 27 07:45:36 crc kubenswrapper[4934]: E1227 07:45:36.770906 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-wtxbn" podUID="b06ee67d-88e4-4c2f-b0a1-efdc7517c40d" Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.871781 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/20da3be4-5db1-43ec-9b86-5ffae5145bf8-secret-volume\") pod \"20da3be4-5db1-43ec-9b86-5ffae5145bf8\" (UID: \"20da3be4-5db1-43ec-9b86-5ffae5145bf8\") " Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.871841 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8skq\" (UniqueName: \"kubernetes.io/projected/20da3be4-5db1-43ec-9b86-5ffae5145bf8-kube-api-access-x8skq\") pod \"20da3be4-5db1-43ec-9b86-5ffae5145bf8\" (UID: \"20da3be4-5db1-43ec-9b86-5ffae5145bf8\") " Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.871876 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwk6g\" (UniqueName: \"kubernetes.io/projected/959b5ca2-d235-4457-98ec-fe077dfcdc6f-kube-api-access-wwk6g\") pod \"959b5ca2-d235-4457-98ec-fe077dfcdc6f\" (UID: \"959b5ca2-d235-4457-98ec-fe077dfcdc6f\") " Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.871904 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20da3be4-5db1-43ec-9b86-5ffae5145bf8-config-volume\") pod \"20da3be4-5db1-43ec-9b86-5ffae5145bf8\" (UID: \"20da3be4-5db1-43ec-9b86-5ffae5145bf8\") " Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.871957 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/959b5ca2-d235-4457-98ec-fe077dfcdc6f-client-ca\") pod \"959b5ca2-d235-4457-98ec-fe077dfcdc6f\" (UID: \"959b5ca2-d235-4457-98ec-fe077dfcdc6f\") " Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.872007 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/959b5ca2-d235-4457-98ec-fe077dfcdc6f-serving-cert\") pod \"959b5ca2-d235-4457-98ec-fe077dfcdc6f\" (UID: \"959b5ca2-d235-4457-98ec-fe077dfcdc6f\") " Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.872032 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/959b5ca2-d235-4457-98ec-fe077dfcdc6f-config\") pod \"959b5ca2-d235-4457-98ec-fe077dfcdc6f\" (UID: \"959b5ca2-d235-4457-98ec-fe077dfcdc6f\") " Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.872743 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20da3be4-5db1-43ec-9b86-5ffae5145bf8-config-volume" (OuterVolumeSpecName: "config-volume") pod "20da3be4-5db1-43ec-9b86-5ffae5145bf8" (UID: "20da3be4-5db1-43ec-9b86-5ffae5145bf8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.873206 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/959b5ca2-d235-4457-98ec-fe077dfcdc6f-config" (OuterVolumeSpecName: "config") pod "959b5ca2-d235-4457-98ec-fe077dfcdc6f" (UID: "959b5ca2-d235-4457-98ec-fe077dfcdc6f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.874047 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/959b5ca2-d235-4457-98ec-fe077dfcdc6f-client-ca" (OuterVolumeSpecName: "client-ca") pod "959b5ca2-d235-4457-98ec-fe077dfcdc6f" (UID: "959b5ca2-d235-4457-98ec-fe077dfcdc6f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.876152 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20da3be4-5db1-43ec-9b86-5ffae5145bf8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "20da3be4-5db1-43ec-9b86-5ffae5145bf8" (UID: "20da3be4-5db1-43ec-9b86-5ffae5145bf8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.876394 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20da3be4-5db1-43ec-9b86-5ffae5145bf8-kube-api-access-x8skq" (OuterVolumeSpecName: "kube-api-access-x8skq") pod "20da3be4-5db1-43ec-9b86-5ffae5145bf8" (UID: "20da3be4-5db1-43ec-9b86-5ffae5145bf8"). InnerVolumeSpecName "kube-api-access-x8skq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.877251 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/959b5ca2-d235-4457-98ec-fe077dfcdc6f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "959b5ca2-d235-4457-98ec-fe077dfcdc6f" (UID: "959b5ca2-d235-4457-98ec-fe077dfcdc6f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.898509 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/959b5ca2-d235-4457-98ec-fe077dfcdc6f-kube-api-access-wwk6g" (OuterVolumeSpecName: "kube-api-access-wwk6g") pod "959b5ca2-d235-4457-98ec-fe077dfcdc6f" (UID: "959b5ca2-d235-4457-98ec-fe077dfcdc6f"). InnerVolumeSpecName "kube-api-access-wwk6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.972979 4934 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20da3be4-5db1-43ec-9b86-5ffae5145bf8-config-volume\") on node \"crc\" DevicePath \"\"" Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.973007 4934 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/959b5ca2-d235-4457-98ec-fe077dfcdc6f-client-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.973017 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/959b5ca2-d235-4457-98ec-fe077dfcdc6f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.973027 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/959b5ca2-d235-4457-98ec-fe077dfcdc6f-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.973036 4934 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/20da3be4-5db1-43ec-9b86-5ffae5145bf8-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.973046 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8skq\" (UniqueName: \"kubernetes.io/projected/20da3be4-5db1-43ec-9b86-5ffae5145bf8-kube-api-access-x8skq\") on node \"crc\" DevicePath \"\"" Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.973058 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwk6g\" (UniqueName: \"kubernetes.io/projected/959b5ca2-d235-4457-98ec-fe077dfcdc6f-kube-api-access-wwk6g\") on node \"crc\" DevicePath \"\"" Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.973272 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" event={"ID":"959b5ca2-d235-4457-98ec-fe077dfcdc6f","Type":"ContainerDied","Data":"151cdaebfd40baaacba6cfe6685f366489e0026332eff9399b0b32cff2895e29"} Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.973312 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc" Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.973333 4934 scope.go:117] "RemoveContainer" containerID="b6f570194919b76e785ecb947ba89694a8df2a47ab393bdaaf2e4cf367a713c1" Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.975339 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs" event={"ID":"20da3be4-5db1-43ec-9b86-5ffae5145bf8","Type":"ContainerDied","Data":"79df9ef1359724aa621eb4d4adf6d48f6ebb9cfbe3ea82ab41ed4b032a1e6f76"} Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.975380 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79df9ef1359724aa621eb4d4adf6d48f6ebb9cfbe3ea82ab41ed4b032a1e6f76" Dec 27 07:45:36 crc kubenswrapper[4934]: I1227 07:45:36.975343 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs" Dec 27 07:45:37 crc kubenswrapper[4934]: I1227 07:45:37.015447 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc"] Dec 27 07:45:37 crc kubenswrapper[4934]: I1227 07:45:37.025570 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-r5zxc"] Dec 27 07:45:37 crc kubenswrapper[4934]: I1227 07:45:37.474795 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="959b5ca2-d235-4457-98ec-fe077dfcdc6f" path="/var/lib/kubelet/pods/959b5ca2-d235-4457-98ec-fe077dfcdc6f/volumes" Dec 27 07:45:37 crc kubenswrapper[4934]: E1227 07:45:37.804809 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 27 07:45:37 crc kubenswrapper[4934]: E1227 07:45:37.805320 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6544m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-gtrfs_openshift-marketplace(c26588b0-e09e-4bac-bfd4-22289100341e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 27 07:45:37 crc kubenswrapper[4934]: E1227 07:45:37.806573 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-gtrfs" podUID="c26588b0-e09e-4bac-bfd4-22289100341e" Dec 27 07:45:38 crc kubenswrapper[4934]: I1227 07:45:38.217037 4934 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-8npds container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 07:45:38 crc kubenswrapper[4934]: I1227 07:45:38.217166 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" podUID="5548c900-e014-4c35-b62e-455cd82f5887" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.571012 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm"] Dec 27 07:45:39 crc kubenswrapper[4934]: E1227 07:45:39.571275 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="959b5ca2-d235-4457-98ec-fe077dfcdc6f" containerName="route-controller-manager" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.571293 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="959b5ca2-d235-4457-98ec-fe077dfcdc6f" containerName="route-controller-manager" Dec 27 07:45:39 crc kubenswrapper[4934]: E1227 07:45:39.571311 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20da3be4-5db1-43ec-9b86-5ffae5145bf8" containerName="collect-profiles" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.571345 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="20da3be4-5db1-43ec-9b86-5ffae5145bf8" containerName="collect-profiles" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.571456 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="20da3be4-5db1-43ec-9b86-5ffae5145bf8" containerName="collect-profiles" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.571471 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="959b5ca2-d235-4457-98ec-fe077dfcdc6f" containerName="route-controller-manager" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.572352 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.574462 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.575164 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.575253 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.575663 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.575964 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.576120 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.595297 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm"] Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.738197 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496c7f69-dbf1-4680-94db-b58b1a0044f2-serving-cert\") pod \"route-controller-manager-796dfb4895-65ztm\" (UID: \"496c7f69-dbf1-4680-94db-b58b1a0044f2\") " pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.738255 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/496c7f69-dbf1-4680-94db-b58b1a0044f2-client-ca\") pod \"route-controller-manager-796dfb4895-65ztm\" (UID: \"496c7f69-dbf1-4680-94db-b58b1a0044f2\") " pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.738281 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgfbr\" (UniqueName: \"kubernetes.io/projected/496c7f69-dbf1-4680-94db-b58b1a0044f2-kube-api-access-mgfbr\") pod \"route-controller-manager-796dfb4895-65ztm\" (UID: \"496c7f69-dbf1-4680-94db-b58b1a0044f2\") " pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.738303 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496c7f69-dbf1-4680-94db-b58b1a0044f2-config\") pod \"route-controller-manager-796dfb4895-65ztm\" (UID: \"496c7f69-dbf1-4680-94db-b58b1a0044f2\") " pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.838975 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496c7f69-dbf1-4680-94db-b58b1a0044f2-config\") pod \"route-controller-manager-796dfb4895-65ztm\" (UID: \"496c7f69-dbf1-4680-94db-b58b1a0044f2\") " pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.839058 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496c7f69-dbf1-4680-94db-b58b1a0044f2-serving-cert\") pod \"route-controller-manager-796dfb4895-65ztm\" (UID: \"496c7f69-dbf1-4680-94db-b58b1a0044f2\") " pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.839111 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/496c7f69-dbf1-4680-94db-b58b1a0044f2-client-ca\") pod \"route-controller-manager-796dfb4895-65ztm\" (UID: \"496c7f69-dbf1-4680-94db-b58b1a0044f2\") " pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.839137 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgfbr\" (UniqueName: \"kubernetes.io/projected/496c7f69-dbf1-4680-94db-b58b1a0044f2-kube-api-access-mgfbr\") pod \"route-controller-manager-796dfb4895-65ztm\" (UID: \"496c7f69-dbf1-4680-94db-b58b1a0044f2\") " pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.840500 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496c7f69-dbf1-4680-94db-b58b1a0044f2-config\") pod \"route-controller-manager-796dfb4895-65ztm\" (UID: \"496c7f69-dbf1-4680-94db-b58b1a0044f2\") " pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.841192 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/496c7f69-dbf1-4680-94db-b58b1a0044f2-client-ca\") pod \"route-controller-manager-796dfb4895-65ztm\" (UID: \"496c7f69-dbf1-4680-94db-b58b1a0044f2\") " pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.845248 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496c7f69-dbf1-4680-94db-b58b1a0044f2-serving-cert\") pod \"route-controller-manager-796dfb4895-65ztm\" (UID: \"496c7f69-dbf1-4680-94db-b58b1a0044f2\") " pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.860974 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgfbr\" (UniqueName: \"kubernetes.io/projected/496c7f69-dbf1-4680-94db-b58b1a0044f2-kube-api-access-mgfbr\") pod \"route-controller-manager-796dfb4895-65ztm\" (UID: \"496c7f69-dbf1-4680-94db-b58b1a0044f2\") " pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" Dec 27 07:45:39 crc kubenswrapper[4934]: I1227 07:45:39.957702 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" Dec 27 07:45:40 crc kubenswrapper[4934]: I1227 07:45:40.039316 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 27 07:45:40 crc kubenswrapper[4934]: I1227 07:45:40.040163 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 27 07:45:40 crc kubenswrapper[4934]: I1227 07:45:40.043874 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 27 07:45:40 crc kubenswrapper[4934]: I1227 07:45:40.143498 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b8ce8d56-fde4-44c6-a3df-53d4dbca22d4-kube-api-access\") pod \"installer-9-crc\" (UID: \"b8ce8d56-fde4-44c6-a3df-53d4dbca22d4\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 27 07:45:40 crc kubenswrapper[4934]: I1227 07:45:40.143625 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b8ce8d56-fde4-44c6-a3df-53d4dbca22d4-var-lock\") pod \"installer-9-crc\" (UID: \"b8ce8d56-fde4-44c6-a3df-53d4dbca22d4\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 27 07:45:40 crc kubenswrapper[4934]: I1227 07:45:40.143673 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b8ce8d56-fde4-44c6-a3df-53d4dbca22d4-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b8ce8d56-fde4-44c6-a3df-53d4dbca22d4\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 27 07:45:40 crc kubenswrapper[4934]: I1227 07:45:40.244639 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b8ce8d56-fde4-44c6-a3df-53d4dbca22d4-kube-api-access\") pod \"installer-9-crc\" (UID: \"b8ce8d56-fde4-44c6-a3df-53d4dbca22d4\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 27 07:45:40 crc kubenswrapper[4934]: I1227 07:45:40.244708 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b8ce8d56-fde4-44c6-a3df-53d4dbca22d4-var-lock\") pod \"installer-9-crc\" (UID: \"b8ce8d56-fde4-44c6-a3df-53d4dbca22d4\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 27 07:45:40 crc kubenswrapper[4934]: I1227 07:45:40.244731 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b8ce8d56-fde4-44c6-a3df-53d4dbca22d4-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b8ce8d56-fde4-44c6-a3df-53d4dbca22d4\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 27 07:45:40 crc kubenswrapper[4934]: I1227 07:45:40.244806 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b8ce8d56-fde4-44c6-a3df-53d4dbca22d4-kubelet-dir\") pod \"installer-9-crc\" (UID: \"b8ce8d56-fde4-44c6-a3df-53d4dbca22d4\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 27 07:45:40 crc kubenswrapper[4934]: I1227 07:45:40.244848 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b8ce8d56-fde4-44c6-a3df-53d4dbca22d4-var-lock\") pod \"installer-9-crc\" (UID: \"b8ce8d56-fde4-44c6-a3df-53d4dbca22d4\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 27 07:45:40 crc kubenswrapper[4934]: I1227 07:45:40.260878 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b8ce8d56-fde4-44c6-a3df-53d4dbca22d4-kube-api-access\") pod \"installer-9-crc\" (UID: \"b8ce8d56-fde4-44c6-a3df-53d4dbca22d4\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 27 07:45:40 crc kubenswrapper[4934]: I1227 07:45:40.363596 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 27 07:45:44 crc kubenswrapper[4934]: E1227 07:45:44.907937 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-wtxbn" podUID="b06ee67d-88e4-4c2f-b0a1-efdc7517c40d" Dec 27 07:45:44 crc kubenswrapper[4934]: E1227 07:45:44.908020 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-gtrfs" podUID="c26588b0-e09e-4bac-bfd4-22289100341e" Dec 27 07:45:45 crc kubenswrapper[4934]: E1227 07:45:45.260417 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 27 07:45:45 crc kubenswrapper[4934]: E1227 07:45:45.260987 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5gltj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-xwb5s_openshift-marketplace(deaa2d15-bcaf-4568-8847-7bccae9205e0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 27 07:45:45 crc kubenswrapper[4934]: E1227 07:45:45.263855 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-xwb5s" podUID="deaa2d15-bcaf-4568-8847-7bccae9205e0" Dec 27 07:45:45 crc kubenswrapper[4934]: E1227 07:45:45.314073 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 27 07:45:45 crc kubenswrapper[4934]: E1227 07:45:45.314312 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zdz7v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-lf8vn_openshift-marketplace(4d8b26f6-cfe5-44d8-89e2-92fff18c71a7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 27 07:45:45 crc kubenswrapper[4934]: E1227 07:45:45.315533 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-lf8vn" podUID="4d8b26f6-cfe5-44d8-89e2-92fff18c71a7" Dec 27 07:45:45 crc kubenswrapper[4934]: I1227 07:45:45.330509 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 07:45:45 crc kubenswrapper[4934]: I1227 07:45:45.330573 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 07:45:45 crc kubenswrapper[4934]: I1227 07:45:45.330625 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 07:45:45 crc kubenswrapper[4934]: I1227 07:45:45.331230 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 07:45:45 crc kubenswrapper[4934]: I1227 07:45:45.331322 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c" gracePeriod=600 Dec 27 07:45:46 crc kubenswrapper[4934]: I1227 07:45:46.037640 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c" exitCode=0 Dec 27 07:45:46 crc kubenswrapper[4934]: I1227 07:45:46.037717 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c"} Dec 27 07:45:46 crc kubenswrapper[4934]: E1227 07:45:46.390055 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-lf8vn" podUID="4d8b26f6-cfe5-44d8-89e2-92fff18c71a7" Dec 27 07:45:46 crc kubenswrapper[4934]: E1227 07:45:46.390301 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-xwb5s" podUID="deaa2d15-bcaf-4568-8847-7bccae9205e0" Dec 27 07:45:46 crc kubenswrapper[4934]: E1227 07:45:46.464165 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 27 07:45:46 crc kubenswrapper[4934]: E1227 07:45:46.464344 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2g8n6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-t4c2w_openshift-marketplace(19b07e13-aa5d-4a0a-b6d9-254afc234025): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 27 07:45:46 crc kubenswrapper[4934]: E1227 07:45:46.465873 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-t4c2w" podUID="19b07e13-aa5d-4a0a-b6d9-254afc234025" Dec 27 07:45:48 crc kubenswrapper[4934]: E1227 07:45:48.131715 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-t4c2w" podUID="19b07e13-aa5d-4a0a-b6d9-254afc234025" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.216913 4934 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-8npds container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.217319 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" podUID="5548c900-e014-4c35-b62e-455cd82f5887" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 07:45:48 crc kubenswrapper[4934]: E1227 07:45:48.217723 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 27 07:45:48 crc kubenswrapper[4934]: E1227 07:45:48.217859 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8ks49,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-5phzx_openshift-marketplace(ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 27 07:45:48 crc kubenswrapper[4934]: E1227 07:45:48.219689 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-5phzx" podUID="ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.223980 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" Dec 27 07:45:48 crc kubenswrapper[4934]: E1227 07:45:48.236883 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 27 07:45:48 crc kubenswrapper[4934]: E1227 07:45:48.237028 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vvtqw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-gk922_openshift-marketplace(5f3aa628-fe89-4de2-ad21-34b7d18c9033): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 27 07:45:48 crc kubenswrapper[4934]: E1227 07:45:48.238317 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-gk922" podUID="5f3aa628-fe89-4de2-ad21-34b7d18c9033" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.264967 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-665c7b5866-brtpg"] Dec 27 07:45:48 crc kubenswrapper[4934]: E1227 07:45:48.265347 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5548c900-e014-4c35-b62e-455cd82f5887" containerName="controller-manager" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.265363 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="5548c900-e014-4c35-b62e-455cd82f5887" containerName="controller-manager" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.265516 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="5548c900-e014-4c35-b62e-455cd82f5887" containerName="controller-manager" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.266055 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.270917 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmk2v\" (UniqueName: \"kubernetes.io/projected/5548c900-e014-4c35-b62e-455cd82f5887-kube-api-access-fmk2v\") pod \"5548c900-e014-4c35-b62e-455cd82f5887\" (UID: \"5548c900-e014-4c35-b62e-455cd82f5887\") " Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.270972 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5548c900-e014-4c35-b62e-455cd82f5887-client-ca\") pod \"5548c900-e014-4c35-b62e-455cd82f5887\" (UID: \"5548c900-e014-4c35-b62e-455cd82f5887\") " Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.271057 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5548c900-e014-4c35-b62e-455cd82f5887-serving-cert\") pod \"5548c900-e014-4c35-b62e-455cd82f5887\" (UID: \"5548c900-e014-4c35-b62e-455cd82f5887\") " Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.271095 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5548c900-e014-4c35-b62e-455cd82f5887-config\") pod \"5548c900-e014-4c35-b62e-455cd82f5887\" (UID: \"5548c900-e014-4c35-b62e-455cd82f5887\") " Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.271138 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5548c900-e014-4c35-b62e-455cd82f5887-proxy-ca-bundles\") pod \"5548c900-e014-4c35-b62e-455cd82f5887\" (UID: \"5548c900-e014-4c35-b62e-455cd82f5887\") " Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.272483 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5548c900-e014-4c35-b62e-455cd82f5887-client-ca" (OuterVolumeSpecName: "client-ca") pod "5548c900-e014-4c35-b62e-455cd82f5887" (UID: "5548c900-e014-4c35-b62e-455cd82f5887"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.273454 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5548c900-e014-4c35-b62e-455cd82f5887-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "5548c900-e014-4c35-b62e-455cd82f5887" (UID: "5548c900-e014-4c35-b62e-455cd82f5887"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.273629 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5548c900-e014-4c35-b62e-455cd82f5887-config" (OuterVolumeSpecName: "config") pod "5548c900-e014-4c35-b62e-455cd82f5887" (UID: "5548c900-e014-4c35-b62e-455cd82f5887"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.275830 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-665c7b5866-brtpg"] Dec 27 07:45:48 crc kubenswrapper[4934]: E1227 07:45:48.278248 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 27 07:45:48 crc kubenswrapper[4934]: E1227 07:45:48.278376 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8qwpc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-lzcth_openshift-marketplace(52a38d21-a579-4f37-be2f-856e2f837f83): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 27 07:45:48 crc kubenswrapper[4934]: E1227 07:45:48.279670 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-lzcth" podUID="52a38d21-a579-4f37-be2f-856e2f837f83" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.302122 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5548c900-e014-4c35-b62e-455cd82f5887-kube-api-access-fmk2v" (OuterVolumeSpecName: "kube-api-access-fmk2v") pod "5548c900-e014-4c35-b62e-455cd82f5887" (UID: "5548c900-e014-4c35-b62e-455cd82f5887"). InnerVolumeSpecName "kube-api-access-fmk2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.317411 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5548c900-e014-4c35-b62e-455cd82f5887-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5548c900-e014-4c35-b62e-455cd82f5887" (UID: "5548c900-e014-4c35-b62e-455cd82f5887"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.372860 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd66a5e0-0214-485f-80dd-0a54e4cec402-config\") pod \"controller-manager-665c7b5866-brtpg\" (UID: \"dd66a5e0-0214-485f-80dd-0a54e4cec402\") " pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.372917 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd66a5e0-0214-485f-80dd-0a54e4cec402-client-ca\") pod \"controller-manager-665c7b5866-brtpg\" (UID: \"dd66a5e0-0214-485f-80dd-0a54e4cec402\") " pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.373027 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd66a5e0-0214-485f-80dd-0a54e4cec402-serving-cert\") pod \"controller-manager-665c7b5866-brtpg\" (UID: \"dd66a5e0-0214-485f-80dd-0a54e4cec402\") " pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.373132 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2r8s\" (UniqueName: \"kubernetes.io/projected/dd66a5e0-0214-485f-80dd-0a54e4cec402-kube-api-access-s2r8s\") pod \"controller-manager-665c7b5866-brtpg\" (UID: \"dd66a5e0-0214-485f-80dd-0a54e4cec402\") " pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.373254 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dd66a5e0-0214-485f-80dd-0a54e4cec402-proxy-ca-bundles\") pod \"controller-manager-665c7b5866-brtpg\" (UID: \"dd66a5e0-0214-485f-80dd-0a54e4cec402\") " pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.373338 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmk2v\" (UniqueName: \"kubernetes.io/projected/5548c900-e014-4c35-b62e-455cd82f5887-kube-api-access-fmk2v\") on node \"crc\" DevicePath \"\"" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.373359 4934 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5548c900-e014-4c35-b62e-455cd82f5887-client-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.373373 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5548c900-e014-4c35-b62e-455cd82f5887-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.373389 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5548c900-e014-4c35-b62e-455cd82f5887-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.373408 4934 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5548c900-e014-4c35-b62e-455cd82f5887-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.474329 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2r8s\" (UniqueName: \"kubernetes.io/projected/dd66a5e0-0214-485f-80dd-0a54e4cec402-kube-api-access-s2r8s\") pod \"controller-manager-665c7b5866-brtpg\" (UID: \"dd66a5e0-0214-485f-80dd-0a54e4cec402\") " pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.474770 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dd66a5e0-0214-485f-80dd-0a54e4cec402-proxy-ca-bundles\") pod \"controller-manager-665c7b5866-brtpg\" (UID: \"dd66a5e0-0214-485f-80dd-0a54e4cec402\") " pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.474835 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd66a5e0-0214-485f-80dd-0a54e4cec402-client-ca\") pod \"controller-manager-665c7b5866-brtpg\" (UID: \"dd66a5e0-0214-485f-80dd-0a54e4cec402\") " pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.474860 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd66a5e0-0214-485f-80dd-0a54e4cec402-config\") pod \"controller-manager-665c7b5866-brtpg\" (UID: \"dd66a5e0-0214-485f-80dd-0a54e4cec402\") " pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.474912 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd66a5e0-0214-485f-80dd-0a54e4cec402-serving-cert\") pod \"controller-manager-665c7b5866-brtpg\" (UID: \"dd66a5e0-0214-485f-80dd-0a54e4cec402\") " pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.476612 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd66a5e0-0214-485f-80dd-0a54e4cec402-config\") pod \"controller-manager-665c7b5866-brtpg\" (UID: \"dd66a5e0-0214-485f-80dd-0a54e4cec402\") " pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.476696 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dd66a5e0-0214-485f-80dd-0a54e4cec402-proxy-ca-bundles\") pod \"controller-manager-665c7b5866-brtpg\" (UID: \"dd66a5e0-0214-485f-80dd-0a54e4cec402\") " pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.477098 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd66a5e0-0214-485f-80dd-0a54e4cec402-client-ca\") pod \"controller-manager-665c7b5866-brtpg\" (UID: \"dd66a5e0-0214-485f-80dd-0a54e4cec402\") " pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.482336 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd66a5e0-0214-485f-80dd-0a54e4cec402-serving-cert\") pod \"controller-manager-665c7b5866-brtpg\" (UID: \"dd66a5e0-0214-485f-80dd-0a54e4cec402\") " pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.491840 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2r8s\" (UniqueName: \"kubernetes.io/projected/dd66a5e0-0214-485f-80dd-0a54e4cec402-kube-api-access-s2r8s\") pod \"controller-manager-665c7b5866-brtpg\" (UID: \"dd66a5e0-0214-485f-80dd-0a54e4cec402\") " pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.517548 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 27 07:45:48 crc kubenswrapper[4934]: W1227 07:45:48.530351 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod729c6c46_3ca1_4b18_abcb_bcbb4fa45cb5.slice/crio-c95f94492b210fe9a8f6d653bcfae17adec076aac185811e9405b300443f2b14 WatchSource:0}: Error finding container c95f94492b210fe9a8f6d653bcfae17adec076aac185811e9405b300443f2b14: Status 404 returned error can't find the container with id c95f94492b210fe9a8f6d653bcfae17adec076aac185811e9405b300443f2b14 Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.660564 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.666990 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm"] Dec 27 07:45:48 crc kubenswrapper[4934]: W1227 07:45:48.673668 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podb8ce8d56_fde4_44c6_a3df_53d4dbca22d4.slice/crio-601e4c6339e39ddf20b8f28c462c006536443819240f41e922def1929c2b11f5 WatchSource:0}: Error finding container 601e4c6339e39ddf20b8f28c462c006536443819240f41e922def1929c2b11f5: Status 404 returned error can't find the container with id 601e4c6339e39ddf20b8f28c462c006536443819240f41e922def1929c2b11f5 Dec 27 07:45:48 crc kubenswrapper[4934]: W1227 07:45:48.676223 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod496c7f69_dbf1_4680_94db_b58b1a0044f2.slice/crio-60cee34f2d024194401e12274e4a5f74c701af376edcd86067713ae94100ec24 WatchSource:0}: Error finding container 60cee34f2d024194401e12274e4a5f74c701af376edcd86067713ae94100ec24: Status 404 returned error can't find the container with id 60cee34f2d024194401e12274e4a5f74c701af376edcd86067713ae94100ec24 Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.676546 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 27 07:45:48 crc kubenswrapper[4934]: I1227 07:45:48.910100 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-665c7b5866-brtpg"] Dec 27 07:45:48 crc kubenswrapper[4934]: W1227 07:45:48.925715 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd66a5e0_0214_485f_80dd_0a54e4cec402.slice/crio-1f1519149193dff3409e9f7ef4b290b4238ee1fd2d2e6a4875ca0daf0bae6a55 WatchSource:0}: Error finding container 1f1519149193dff3409e9f7ef4b290b4238ee1fd2d2e6a4875ca0daf0bae6a55: Status 404 returned error can't find the container with id 1f1519149193dff3409e9f7ef4b290b4238ee1fd2d2e6a4875ca0daf0bae6a55 Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.058610 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"44124e136c1edec767fa7b55e82033376dbd4c05224d1c80d3b03b40d50a5a60"} Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.060119 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" event={"ID":"dd66a5e0-0214-485f-80dd-0a54e4cec402","Type":"ContainerStarted","Data":"806d607c93d8584bd0c4554a5b4d3ddf7e88a2f57ae29a6b60025905f4d866f4"} Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.060875 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" event={"ID":"dd66a5e0-0214-485f-80dd-0a54e4cec402","Type":"ContainerStarted","Data":"1f1519149193dff3409e9f7ef4b290b4238ee1fd2d2e6a4875ca0daf0bae6a55"} Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.060989 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.063041 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" event={"ID":"5548c900-e014-4c35-b62e-455cd82f5887","Type":"ContainerDied","Data":"fa0e38b01db058966e64f388fec75289e82c54449b95f88ee4b3b3c16f2a8db3"} Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.063075 4934 scope.go:117] "RemoveContainer" containerID="c59d13363d1636e34f915d67816d51af03491a24eef81601c054e63b8657f516" Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.063463 4934 patch_prober.go:28] interesting pod/controller-manager-665c7b5866-brtpg container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" start-of-body= Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.063504 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" podUID="dd66a5e0-0214-485f-80dd-0a54e4cec402" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": dial tcp 10.217.0.58:8443: connect: connection refused" Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.063831 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-8npds" Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.064949 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5","Type":"ContainerStarted","Data":"078d5207b1a94aa15b8666be6b67dd6d1c874732028d30715d324c56f0aa258f"} Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.065065 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5","Type":"ContainerStarted","Data":"c95f94492b210fe9a8f6d653bcfae17adec076aac185811e9405b300443f2b14"} Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.072932 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" event={"ID":"496c7f69-dbf1-4680-94db-b58b1a0044f2","Type":"ContainerStarted","Data":"79a560e21aded6e113b938f5956cda93ad5a2949534d477f0d1ba72d62f91631"} Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.072972 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" event={"ID":"496c7f69-dbf1-4680-94db-b58b1a0044f2","Type":"ContainerStarted","Data":"60cee34f2d024194401e12274e4a5f74c701af376edcd86067713ae94100ec24"} Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.073445 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.076727 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b8ce8d56-fde4-44c6-a3df-53d4dbca22d4","Type":"ContainerStarted","Data":"1321590a3c6e68643a10b3d4b7e2a2ec45de9e3034e965e3b58c609d93c3d3a3"} Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.076855 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b8ce8d56-fde4-44c6-a3df-53d4dbca22d4","Type":"ContainerStarted","Data":"601e4c6339e39ddf20b8f28c462c006536443819240f41e922def1929c2b11f5"} Dec 27 07:45:49 crc kubenswrapper[4934]: E1227 07:45:49.087152 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-lzcth" podUID="52a38d21-a579-4f37-be2f-856e2f837f83" Dec 27 07:45:49 crc kubenswrapper[4934]: E1227 07:45:49.094866 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-gk922" podUID="5f3aa628-fe89-4de2-ad21-34b7d18c9033" Dec 27 07:45:49 crc kubenswrapper[4934]: E1227 07:45:49.094956 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5phzx" podUID="ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.177811 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" podStartSLOduration=20.177783377 podStartE2EDuration="20.177783377s" podCreationTimestamp="2025-12-27 07:45:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:45:49.172821981 +0000 UTC m=+209.993262575" watchObservedRunningTime="2025-12-27 07:45:49.177783377 +0000 UTC m=+209.998223981" Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.186022 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=9.18600388 podStartE2EDuration="9.18600388s" podCreationTimestamp="2025-12-27 07:45:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:45:49.185969789 +0000 UTC m=+210.006410403" watchObservedRunningTime="2025-12-27 07:45:49.18600388 +0000 UTC m=+210.006444474" Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.205586 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=15.205565847 podStartE2EDuration="15.205565847s" podCreationTimestamp="2025-12-27 07:45:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:45:49.201567129 +0000 UTC m=+210.022007733" watchObservedRunningTime="2025-12-27 07:45:49.205565847 +0000 UTC m=+210.026006441" Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.224169 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" podStartSLOduration=20.224144165 podStartE2EDuration="20.224144165s" podCreationTimestamp="2025-12-27 07:45:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:45:49.221688752 +0000 UTC m=+210.042129346" watchObservedRunningTime="2025-12-27 07:45:49.224144165 +0000 UTC m=+210.044584779" Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.234271 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8npds"] Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.239274 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-8npds"] Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.375655 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" Dec 27 07:45:49 crc kubenswrapper[4934]: I1227 07:45:49.474222 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5548c900-e014-4c35-b62e-455cd82f5887" path="/var/lib/kubelet/pods/5548c900-e014-4c35-b62e-455cd82f5887/volumes" Dec 27 07:45:50 crc kubenswrapper[4934]: I1227 07:45:50.083411 4934 generic.go:334] "Generic (PLEG): container finished" podID="729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5" containerID="078d5207b1a94aa15b8666be6b67dd6d1c874732028d30715d324c56f0aa258f" exitCode=0 Dec 27 07:45:50 crc kubenswrapper[4934]: I1227 07:45:50.083522 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5","Type":"ContainerDied","Data":"078d5207b1a94aa15b8666be6b67dd6d1c874732028d30715d324c56f0aa258f"} Dec 27 07:45:50 crc kubenswrapper[4934]: I1227 07:45:50.092927 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" Dec 27 07:45:51 crc kubenswrapper[4934]: I1227 07:45:51.430788 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 27 07:45:51 crc kubenswrapper[4934]: I1227 07:45:51.518566 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5-kubelet-dir\") pod \"729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5\" (UID: \"729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5\") " Dec 27 07:45:51 crc kubenswrapper[4934]: I1227 07:45:51.519051 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5-kube-api-access\") pod \"729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5\" (UID: \"729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5\") " Dec 27 07:45:51 crc kubenswrapper[4934]: I1227 07:45:51.518759 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5" (UID: "729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:45:51 crc kubenswrapper[4934]: I1227 07:45:51.519412 4934 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 27 07:45:51 crc kubenswrapper[4934]: I1227 07:45:51.525585 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5" (UID: "729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:45:51 crc kubenswrapper[4934]: I1227 07:45:51.620244 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 27 07:45:52 crc kubenswrapper[4934]: I1227 07:45:52.102995 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5","Type":"ContainerDied","Data":"c95f94492b210fe9a8f6d653bcfae17adec076aac185811e9405b300443f2b14"} Dec 27 07:45:52 crc kubenswrapper[4934]: I1227 07:45:52.103050 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 27 07:45:52 crc kubenswrapper[4934]: I1227 07:45:52.103070 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c95f94492b210fe9a8f6d653bcfae17adec076aac185811e9405b300443f2b14" Dec 27 07:45:59 crc kubenswrapper[4934]: I1227 07:45:59.151846 4934 generic.go:334] "Generic (PLEG): container finished" podID="b06ee67d-88e4-4c2f-b0a1-efdc7517c40d" containerID="71b561737dc74ba67d79f3d9f01f0d37d5d880a14a08a55d05dd373992359eaf" exitCode=0 Dec 27 07:45:59 crc kubenswrapper[4934]: I1227 07:45:59.151925 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wtxbn" event={"ID":"b06ee67d-88e4-4c2f-b0a1-efdc7517c40d","Type":"ContainerDied","Data":"71b561737dc74ba67d79f3d9f01f0d37d5d880a14a08a55d05dd373992359eaf"} Dec 27 07:45:59 crc kubenswrapper[4934]: I1227 07:45:59.156735 4934 generic.go:334] "Generic (PLEG): container finished" podID="c26588b0-e09e-4bac-bfd4-22289100341e" containerID="ef3d8fcc3134c47c91d6bf06abe176a1d3915e3bd245d03b5e91b31d9082697e" exitCode=0 Dec 27 07:45:59 crc kubenswrapper[4934]: I1227 07:45:59.156772 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtrfs" event={"ID":"c26588b0-e09e-4bac-bfd4-22289100341e","Type":"ContainerDied","Data":"ef3d8fcc3134c47c91d6bf06abe176a1d3915e3bd245d03b5e91b31d9082697e"} Dec 27 07:46:00 crc kubenswrapper[4934]: I1227 07:46:00.163171 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtrfs" event={"ID":"c26588b0-e09e-4bac-bfd4-22289100341e","Type":"ContainerStarted","Data":"c9cfec9cae89d6648484ad3b175a8774da0551baa5946d59da5555273e9fac61"} Dec 27 07:46:00 crc kubenswrapper[4934]: I1227 07:46:00.165002 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wtxbn" event={"ID":"b06ee67d-88e4-4c2f-b0a1-efdc7517c40d","Type":"ContainerStarted","Data":"0cfd04443627ca28e845e4c8ca2baec3709c700cef244dd8d213200cb5e64c1a"} Dec 27 07:46:00 crc kubenswrapper[4934]: I1227 07:46:00.180148 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gtrfs" podStartSLOduration=5.057270267 podStartE2EDuration="1m10.180127663s" podCreationTimestamp="2025-12-27 07:44:50 +0000 UTC" firstStartedPulling="2025-12-27 07:44:54.432539872 +0000 UTC m=+155.252980466" lastFinishedPulling="2025-12-27 07:45:59.555397268 +0000 UTC m=+220.375837862" observedRunningTime="2025-12-27 07:46:00.176553518 +0000 UTC m=+220.996994132" watchObservedRunningTime="2025-12-27 07:46:00.180127663 +0000 UTC m=+221.000568257" Dec 27 07:46:00 crc kubenswrapper[4934]: I1227 07:46:00.199150 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wtxbn" podStartSLOduration=5.137200085 podStartE2EDuration="1m10.199134244s" podCreationTimestamp="2025-12-27 07:44:50 +0000 UTC" firstStartedPulling="2025-12-27 07:44:54.482341041 +0000 UTC m=+155.302781635" lastFinishedPulling="2025-12-27 07:45:59.5442752 +0000 UTC m=+220.364715794" observedRunningTime="2025-12-27 07:46:00.197277499 +0000 UTC m=+221.017718103" watchObservedRunningTime="2025-12-27 07:46:00.199134244 +0000 UTC m=+221.019574838" Dec 27 07:46:01 crc kubenswrapper[4934]: I1227 07:46:01.015717 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wtxbn" Dec 27 07:46:01 crc kubenswrapper[4934]: I1227 07:46:01.015768 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wtxbn" Dec 27 07:46:01 crc kubenswrapper[4934]: I1227 07:46:01.394723 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gtrfs" Dec 27 07:46:01 crc kubenswrapper[4934]: I1227 07:46:01.395073 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gtrfs" Dec 27 07:46:02 crc kubenswrapper[4934]: I1227 07:46:02.515615 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-wtxbn" podUID="b06ee67d-88e4-4c2f-b0a1-efdc7517c40d" containerName="registry-server" probeResult="failure" output=< Dec 27 07:46:02 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 07:46:02 crc kubenswrapper[4934]: > Dec 27 07:46:02 crc kubenswrapper[4934]: I1227 07:46:02.518378 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-gtrfs" podUID="c26588b0-e09e-4bac-bfd4-22289100341e" containerName="registry-server" probeResult="failure" output=< Dec 27 07:46:02 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 07:46:02 crc kubenswrapper[4934]: > Dec 27 07:46:03 crc kubenswrapper[4934]: I1227 07:46:03.192257 4934 generic.go:334] "Generic (PLEG): container finished" podID="4d8b26f6-cfe5-44d8-89e2-92fff18c71a7" containerID="d3b81630a2af7e42360f9a05fd67fd96cc75605a7b5e7cce2fed372b0504bd27" exitCode=0 Dec 27 07:46:03 crc kubenswrapper[4934]: I1227 07:46:03.192323 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lf8vn" event={"ID":"4d8b26f6-cfe5-44d8-89e2-92fff18c71a7","Type":"ContainerDied","Data":"d3b81630a2af7e42360f9a05fd67fd96cc75605a7b5e7cce2fed372b0504bd27"} Dec 27 07:46:03 crc kubenswrapper[4934]: I1227 07:46:03.197592 4934 generic.go:334] "Generic (PLEG): container finished" podID="19b07e13-aa5d-4a0a-b6d9-254afc234025" containerID="d2c75cb43e3c4cebb027bfd283f92d47af8269e44e26f182db86e42ed276a7d9" exitCode=0 Dec 27 07:46:03 crc kubenswrapper[4934]: I1227 07:46:03.197693 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t4c2w" event={"ID":"19b07e13-aa5d-4a0a-b6d9-254afc234025","Type":"ContainerDied","Data":"d2c75cb43e3c4cebb027bfd283f92d47af8269e44e26f182db86e42ed276a7d9"} Dec 27 07:46:03 crc kubenswrapper[4934]: I1227 07:46:03.201395 4934 generic.go:334] "Generic (PLEG): container finished" podID="deaa2d15-bcaf-4568-8847-7bccae9205e0" containerID="d309456782220cc84313b79fc648c59070da988f5b969378b2b9f9572c4703c8" exitCode=0 Dec 27 07:46:03 crc kubenswrapper[4934]: I1227 07:46:03.201430 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwb5s" event={"ID":"deaa2d15-bcaf-4568-8847-7bccae9205e0","Type":"ContainerDied","Data":"d309456782220cc84313b79fc648c59070da988f5b969378b2b9f9572c4703c8"} Dec 27 07:46:04 crc kubenswrapper[4934]: I1227 07:46:04.216685 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwb5s" event={"ID":"deaa2d15-bcaf-4568-8847-7bccae9205e0","Type":"ContainerStarted","Data":"d0b8d1737acd7fb9e25b1e3efe0bd4d63fd177cc8ee6686b79c81a578ffda4e0"} Dec 27 07:46:04 crc kubenswrapper[4934]: I1227 07:46:04.219289 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lf8vn" event={"ID":"4d8b26f6-cfe5-44d8-89e2-92fff18c71a7","Type":"ContainerStarted","Data":"4ce2add7f5e513f5a9c3021049bc64929393929481adb6ccfe7765397a5fa8a4"} Dec 27 07:46:04 crc kubenswrapper[4934]: I1227 07:46:04.220974 4934 generic.go:334] "Generic (PLEG): container finished" podID="5f3aa628-fe89-4de2-ad21-34b7d18c9033" containerID="4b1029c1f10283da6f1ef020f2ae506a5768e782b079e1e0accbb92ed4c3c21d" exitCode=0 Dec 27 07:46:04 crc kubenswrapper[4934]: I1227 07:46:04.221020 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gk922" event={"ID":"5f3aa628-fe89-4de2-ad21-34b7d18c9033","Type":"ContainerDied","Data":"4b1029c1f10283da6f1ef020f2ae506a5768e782b079e1e0accbb92ed4c3c21d"} Dec 27 07:46:04 crc kubenswrapper[4934]: I1227 07:46:04.225055 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t4c2w" event={"ID":"19b07e13-aa5d-4a0a-b6d9-254afc234025","Type":"ContainerStarted","Data":"682e5c35be37c30ea824e8afa399bbc5077d2e5832bcf98b79ff6d0c038ecf16"} Dec 27 07:46:05 crc kubenswrapper[4934]: I1227 07:46:05.238096 4934 generic.go:334] "Generic (PLEG): container finished" podID="ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" containerID="27f3b35dbbee84864846f9a6d9d209cfdb9d2e5bed7d086af8dd62c00a54e841" exitCode=0 Dec 27 07:46:05 crc kubenswrapper[4934]: I1227 07:46:05.238120 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5phzx" event={"ID":"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20","Type":"ContainerDied","Data":"27f3b35dbbee84864846f9a6d9d209cfdb9d2e5bed7d086af8dd62c00a54e841"} Dec 27 07:46:05 crc kubenswrapper[4934]: I1227 07:46:05.240046 4934 generic.go:334] "Generic (PLEG): container finished" podID="52a38d21-a579-4f37-be2f-856e2f837f83" containerID="8ff79859d5acf8d7d0c0f5eb1aa4bdf590d7a34bcd57d350811631c01fff9670" exitCode=0 Dec 27 07:46:05 crc kubenswrapper[4934]: I1227 07:46:05.240165 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lzcth" event={"ID":"52a38d21-a579-4f37-be2f-856e2f837f83","Type":"ContainerDied","Data":"8ff79859d5acf8d7d0c0f5eb1aa4bdf590d7a34bcd57d350811631c01fff9670"} Dec 27 07:46:05 crc kubenswrapper[4934]: I1227 07:46:05.268410 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xwb5s" podStartSLOduration=4.063836526 podStartE2EDuration="1m12.268394679s" podCreationTimestamp="2025-12-27 07:44:53 +0000 UTC" firstStartedPulling="2025-12-27 07:44:55.527665 +0000 UTC m=+156.348105594" lastFinishedPulling="2025-12-27 07:46:03.732223123 +0000 UTC m=+224.552663747" observedRunningTime="2025-12-27 07:46:05.267318817 +0000 UTC m=+226.087759411" watchObservedRunningTime="2025-12-27 07:46:05.268394679 +0000 UTC m=+226.088835273" Dec 27 07:46:05 crc kubenswrapper[4934]: I1227 07:46:05.305343 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t4c2w" podStartSLOduration=3.966525723 podStartE2EDuration="1m13.305320688s" podCreationTimestamp="2025-12-27 07:44:52 +0000 UTC" firstStartedPulling="2025-12-27 07:44:54.418878319 +0000 UTC m=+155.239318913" lastFinishedPulling="2025-12-27 07:46:03.757673284 +0000 UTC m=+224.578113878" observedRunningTime="2025-12-27 07:46:05.304593267 +0000 UTC m=+226.125033871" watchObservedRunningTime="2025-12-27 07:46:05.305320688 +0000 UTC m=+226.125761282" Dec 27 07:46:05 crc kubenswrapper[4934]: I1227 07:46:05.321593 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lf8vn" podStartSLOduration=4.07513482 podStartE2EDuration="1m12.321578628s" podCreationTimestamp="2025-12-27 07:44:53 +0000 UTC" firstStartedPulling="2025-12-27 07:44:55.554205733 +0000 UTC m=+156.374646327" lastFinishedPulling="2025-12-27 07:46:03.800649521 +0000 UTC m=+224.621090135" observedRunningTime="2025-12-27 07:46:05.319584689 +0000 UTC m=+226.140025303" watchObservedRunningTime="2025-12-27 07:46:05.321578628 +0000 UTC m=+226.142019222" Dec 27 07:46:06 crc kubenswrapper[4934]: I1227 07:46:06.247609 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gk922" event={"ID":"5f3aa628-fe89-4de2-ad21-34b7d18c9033","Type":"ContainerStarted","Data":"09dbaed5045f6892fb85182c594fb8fdb229669203a70d9713f40e7f7afe6dca"} Dec 27 07:46:06 crc kubenswrapper[4934]: I1227 07:46:06.263835 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gk922" podStartSLOduration=4.5743894560000005 podStartE2EDuration="1m16.263819206s" podCreationTimestamp="2025-12-27 07:44:50 +0000 UTC" firstStartedPulling="2025-12-27 07:44:53.326695809 +0000 UTC m=+154.147136403" lastFinishedPulling="2025-12-27 07:46:05.016125559 +0000 UTC m=+225.836566153" observedRunningTime="2025-12-27 07:46:06.261863198 +0000 UTC m=+227.082303792" watchObservedRunningTime="2025-12-27 07:46:06.263819206 +0000 UTC m=+227.084259800" Dec 27 07:46:09 crc kubenswrapper[4934]: I1227 07:46:09.742629 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-665c7b5866-brtpg"] Dec 27 07:46:09 crc kubenswrapper[4934]: I1227 07:46:09.743173 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" podUID="dd66a5e0-0214-485f-80dd-0a54e4cec402" containerName="controller-manager" containerID="cri-o://806d607c93d8584bd0c4554a5b4d3ddf7e88a2f57ae29a6b60025905f4d866f4" gracePeriod=30 Dec 27 07:46:09 crc kubenswrapper[4934]: I1227 07:46:09.839949 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm"] Dec 27 07:46:09 crc kubenswrapper[4934]: I1227 07:46:09.840199 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" podUID="496c7f69-dbf1-4680-94db-b58b1a0044f2" containerName="route-controller-manager" containerID="cri-o://79a560e21aded6e113b938f5956cda93ad5a2949534d477f0d1ba72d62f91631" gracePeriod=30 Dec 27 07:46:09 crc kubenswrapper[4934]: I1227 07:46:09.958364 4934 patch_prober.go:28] interesting pod/route-controller-manager-796dfb4895-65ztm container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" start-of-body= Dec 27 07:46:09 crc kubenswrapper[4934]: I1227 07:46:09.958708 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" podUID="496c7f69-dbf1-4680-94db-b58b1a0044f2" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" Dec 27 07:46:11 crc kubenswrapper[4934]: I1227 07:46:11.310390 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gk922" Dec 27 07:46:11 crc kubenswrapper[4934]: I1227 07:46:11.310461 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gk922" Dec 27 07:46:11 crc kubenswrapper[4934]: I1227 07:46:11.856832 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gk922" Dec 27 07:46:11 crc kubenswrapper[4934]: I1227 07:46:11.857251 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wtxbn" Dec 27 07:46:11 crc kubenswrapper[4934]: I1227 07:46:11.858722 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gtrfs" Dec 27 07:46:11 crc kubenswrapper[4934]: I1227 07:46:11.905713 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gtrfs" Dec 27 07:46:11 crc kubenswrapper[4934]: I1227 07:46:11.913755 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wtxbn" Dec 27 07:46:12 crc kubenswrapper[4934]: I1227 07:46:12.358266 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gk922" Dec 27 07:46:12 crc kubenswrapper[4934]: I1227 07:46:12.897724 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gtrfs"] Dec 27 07:46:12 crc kubenswrapper[4934]: I1227 07:46:12.952742 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t4c2w" Dec 27 07:46:12 crc kubenswrapper[4934]: I1227 07:46:12.953604 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t4c2w" Dec 27 07:46:12 crc kubenswrapper[4934]: I1227 07:46:12.994727 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t4c2w" Dec 27 07:46:13 crc kubenswrapper[4934]: I1227 07:46:13.300844 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gtrfs" podUID="c26588b0-e09e-4bac-bfd4-22289100341e" containerName="registry-server" containerID="cri-o://c9cfec9cae89d6648484ad3b175a8774da0551baa5946d59da5555273e9fac61" gracePeriod=2 Dec 27 07:46:13 crc kubenswrapper[4934]: I1227 07:46:13.379388 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t4c2w" Dec 27 07:46:13 crc kubenswrapper[4934]: I1227 07:46:13.949645 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xwb5s" Dec 27 07:46:13 crc kubenswrapper[4934]: I1227 07:46:13.949990 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xwb5s" Dec 27 07:46:13 crc kubenswrapper[4934]: I1227 07:46:13.997141 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xwb5s" Dec 27 07:46:14 crc kubenswrapper[4934]: I1227 07:46:14.311876 4934 generic.go:334] "Generic (PLEG): container finished" podID="dd66a5e0-0214-485f-80dd-0a54e4cec402" containerID="806d607c93d8584bd0c4554a5b4d3ddf7e88a2f57ae29a6b60025905f4d866f4" exitCode=0 Dec 27 07:46:14 crc kubenswrapper[4934]: I1227 07:46:14.312012 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" event={"ID":"dd66a5e0-0214-485f-80dd-0a54e4cec402","Type":"ContainerDied","Data":"806d607c93d8584bd0c4554a5b4d3ddf7e88a2f57ae29a6b60025905f4d866f4"} Dec 27 07:46:14 crc kubenswrapper[4934]: I1227 07:46:14.351769 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lf8vn" Dec 27 07:46:14 crc kubenswrapper[4934]: I1227 07:46:14.351829 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lf8vn" Dec 27 07:46:14 crc kubenswrapper[4934]: I1227 07:46:14.407876 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lf8vn" Dec 27 07:46:14 crc kubenswrapper[4934]: I1227 07:46:14.700838 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gk922"] Dec 27 07:46:16 crc kubenswrapper[4934]: I1227 07:46:16.149126 4934 generic.go:334] "Generic (PLEG): container finished" podID="496c7f69-dbf1-4680-94db-b58b1a0044f2" containerID="79a560e21aded6e113b938f5956cda93ad5a2949534d477f0d1ba72d62f91631" exitCode=0 Dec 27 07:46:16 crc kubenswrapper[4934]: I1227 07:46:16.149777 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" event={"ID":"496c7f69-dbf1-4680-94db-b58b1a0044f2","Type":"ContainerDied","Data":"79a560e21aded6e113b938f5956cda93ad5a2949534d477f0d1ba72d62f91631"} Dec 27 07:46:16 crc kubenswrapper[4934]: I1227 07:46:16.150424 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gk922" podUID="5f3aa628-fe89-4de2-ad21-34b7d18c9033" containerName="registry-server" containerID="cri-o://09dbaed5045f6892fb85182c594fb8fdb229669203a70d9713f40e7f7afe6dca" gracePeriod=2 Dec 27 07:46:16 crc kubenswrapper[4934]: I1227 07:46:16.218981 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xwb5s" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.161032 4934 generic.go:334] "Generic (PLEG): container finished" podID="5f3aa628-fe89-4de2-ad21-34b7d18c9033" containerID="09dbaed5045f6892fb85182c594fb8fdb229669203a70d9713f40e7f7afe6dca" exitCode=0 Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.161279 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gk922" event={"ID":"5f3aa628-fe89-4de2-ad21-34b7d18c9033","Type":"ContainerDied","Data":"09dbaed5045f6892fb85182c594fb8fdb229669203a70d9713f40e7f7afe6dca"} Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.164735 4934 generic.go:334] "Generic (PLEG): container finished" podID="c26588b0-e09e-4bac-bfd4-22289100341e" containerID="c9cfec9cae89d6648484ad3b175a8774da0551baa5946d59da5555273e9fac61" exitCode=0 Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.165860 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtrfs" event={"ID":"c26588b0-e09e-4bac-bfd4-22289100341e","Type":"ContainerDied","Data":"c9cfec9cae89d6648484ad3b175a8774da0551baa5946d59da5555273e9fac61"} Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.221473 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lf8vn" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.534440 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.543537 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.571284 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj"] Dec 27 07:46:17 crc kubenswrapper[4934]: E1227 07:46:17.571516 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5" containerName="pruner" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.571529 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5" containerName="pruner" Dec 27 07:46:17 crc kubenswrapper[4934]: E1227 07:46:17.571541 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="496c7f69-dbf1-4680-94db-b58b1a0044f2" containerName="route-controller-manager" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.571547 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="496c7f69-dbf1-4680-94db-b58b1a0044f2" containerName="route-controller-manager" Dec 27 07:46:17 crc kubenswrapper[4934]: E1227 07:46:17.571557 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd66a5e0-0214-485f-80dd-0a54e4cec402" containerName="controller-manager" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.571564 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd66a5e0-0214-485f-80dd-0a54e4cec402" containerName="controller-manager" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.571685 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="496c7f69-dbf1-4680-94db-b58b1a0044f2" containerName="route-controller-manager" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.571703 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="729c6c46-3ca1-4b18-abcb-bcbb4fa45cb5" containerName="pruner" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.571713 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd66a5e0-0214-485f-80dd-0a54e4cec402" containerName="controller-manager" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.572094 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.580652 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj"] Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.632116 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dd66a5e0-0214-485f-80dd-0a54e4cec402-proxy-ca-bundles\") pod \"dd66a5e0-0214-485f-80dd-0a54e4cec402\" (UID: \"dd66a5e0-0214-485f-80dd-0a54e4cec402\") " Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.632171 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd66a5e0-0214-485f-80dd-0a54e4cec402-serving-cert\") pod \"dd66a5e0-0214-485f-80dd-0a54e4cec402\" (UID: \"dd66a5e0-0214-485f-80dd-0a54e4cec402\") " Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.632196 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd66a5e0-0214-485f-80dd-0a54e4cec402-config\") pod \"dd66a5e0-0214-485f-80dd-0a54e4cec402\" (UID: \"dd66a5e0-0214-485f-80dd-0a54e4cec402\") " Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.632230 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgfbr\" (UniqueName: \"kubernetes.io/projected/496c7f69-dbf1-4680-94db-b58b1a0044f2-kube-api-access-mgfbr\") pod \"496c7f69-dbf1-4680-94db-b58b1a0044f2\" (UID: \"496c7f69-dbf1-4680-94db-b58b1a0044f2\") " Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.632249 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496c7f69-dbf1-4680-94db-b58b1a0044f2-serving-cert\") pod \"496c7f69-dbf1-4680-94db-b58b1a0044f2\" (UID: \"496c7f69-dbf1-4680-94db-b58b1a0044f2\") " Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.632278 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/496c7f69-dbf1-4680-94db-b58b1a0044f2-client-ca\") pod \"496c7f69-dbf1-4680-94db-b58b1a0044f2\" (UID: \"496c7f69-dbf1-4680-94db-b58b1a0044f2\") " Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.632333 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2r8s\" (UniqueName: \"kubernetes.io/projected/dd66a5e0-0214-485f-80dd-0a54e4cec402-kube-api-access-s2r8s\") pod \"dd66a5e0-0214-485f-80dd-0a54e4cec402\" (UID: \"dd66a5e0-0214-485f-80dd-0a54e4cec402\") " Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.632399 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496c7f69-dbf1-4680-94db-b58b1a0044f2-config\") pod \"496c7f69-dbf1-4680-94db-b58b1a0044f2\" (UID: \"496c7f69-dbf1-4680-94db-b58b1a0044f2\") " Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.632436 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd66a5e0-0214-485f-80dd-0a54e4cec402-client-ca\") pod \"dd66a5e0-0214-485f-80dd-0a54e4cec402\" (UID: \"dd66a5e0-0214-485f-80dd-0a54e4cec402\") " Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.632596 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f89e5ac0-bc03-4af9-91bb-918412fa9b47-config\") pod \"route-controller-manager-58564d585f-nmxfj\" (UID: \"f89e5ac0-bc03-4af9-91bb-918412fa9b47\") " pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.632621 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f89e5ac0-bc03-4af9-91bb-918412fa9b47-serving-cert\") pod \"route-controller-manager-58564d585f-nmxfj\" (UID: \"f89e5ac0-bc03-4af9-91bb-918412fa9b47\") " pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.632665 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72bd5\" (UniqueName: \"kubernetes.io/projected/f89e5ac0-bc03-4af9-91bb-918412fa9b47-kube-api-access-72bd5\") pod \"route-controller-manager-58564d585f-nmxfj\" (UID: \"f89e5ac0-bc03-4af9-91bb-918412fa9b47\") " pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.632701 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f89e5ac0-bc03-4af9-91bb-918412fa9b47-client-ca\") pod \"route-controller-manager-58564d585f-nmxfj\" (UID: \"f89e5ac0-bc03-4af9-91bb-918412fa9b47\") " pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.633487 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd66a5e0-0214-485f-80dd-0a54e4cec402-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "dd66a5e0-0214-485f-80dd-0a54e4cec402" (UID: "dd66a5e0-0214-485f-80dd-0a54e4cec402"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.633934 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496c7f69-dbf1-4680-94db-b58b1a0044f2-client-ca" (OuterVolumeSpecName: "client-ca") pod "496c7f69-dbf1-4680-94db-b58b1a0044f2" (UID: "496c7f69-dbf1-4680-94db-b58b1a0044f2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.634564 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd66a5e0-0214-485f-80dd-0a54e4cec402-config" (OuterVolumeSpecName: "config") pod "dd66a5e0-0214-485f-80dd-0a54e4cec402" (UID: "dd66a5e0-0214-485f-80dd-0a54e4cec402"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.635836 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd66a5e0-0214-485f-80dd-0a54e4cec402-client-ca" (OuterVolumeSpecName: "client-ca") pod "dd66a5e0-0214-485f-80dd-0a54e4cec402" (UID: "dd66a5e0-0214-485f-80dd-0a54e4cec402"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.638978 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496c7f69-dbf1-4680-94db-b58b1a0044f2-kube-api-access-mgfbr" (OuterVolumeSpecName: "kube-api-access-mgfbr") pod "496c7f69-dbf1-4680-94db-b58b1a0044f2" (UID: "496c7f69-dbf1-4680-94db-b58b1a0044f2"). InnerVolumeSpecName "kube-api-access-mgfbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.639254 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd66a5e0-0214-485f-80dd-0a54e4cec402-kube-api-access-s2r8s" (OuterVolumeSpecName: "kube-api-access-s2r8s") pod "dd66a5e0-0214-485f-80dd-0a54e4cec402" (UID: "dd66a5e0-0214-485f-80dd-0a54e4cec402"). InnerVolumeSpecName "kube-api-access-s2r8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.639586 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd66a5e0-0214-485f-80dd-0a54e4cec402-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "dd66a5e0-0214-485f-80dd-0a54e4cec402" (UID: "dd66a5e0-0214-485f-80dd-0a54e4cec402"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.639782 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496c7f69-dbf1-4680-94db-b58b1a0044f2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496c7f69-dbf1-4680-94db-b58b1a0044f2" (UID: "496c7f69-dbf1-4680-94db-b58b1a0044f2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.672329 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496c7f69-dbf1-4680-94db-b58b1a0044f2-config" (OuterVolumeSpecName: "config") pod "496c7f69-dbf1-4680-94db-b58b1a0044f2" (UID: "496c7f69-dbf1-4680-94db-b58b1a0044f2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.695558 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lf8vn"] Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.733502 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f89e5ac0-bc03-4af9-91bb-918412fa9b47-config\") pod \"route-controller-manager-58564d585f-nmxfj\" (UID: \"f89e5ac0-bc03-4af9-91bb-918412fa9b47\") " pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.733561 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f89e5ac0-bc03-4af9-91bb-918412fa9b47-serving-cert\") pod \"route-controller-manager-58564d585f-nmxfj\" (UID: \"f89e5ac0-bc03-4af9-91bb-918412fa9b47\") " pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.733621 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72bd5\" (UniqueName: \"kubernetes.io/projected/f89e5ac0-bc03-4af9-91bb-918412fa9b47-kube-api-access-72bd5\") pod \"route-controller-manager-58564d585f-nmxfj\" (UID: \"f89e5ac0-bc03-4af9-91bb-918412fa9b47\") " pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.733668 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f89e5ac0-bc03-4af9-91bb-918412fa9b47-client-ca\") pod \"route-controller-manager-58564d585f-nmxfj\" (UID: \"f89e5ac0-bc03-4af9-91bb-918412fa9b47\") " pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.733729 4934 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dd66a5e0-0214-485f-80dd-0a54e4cec402-client-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.733760 4934 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dd66a5e0-0214-485f-80dd-0a54e4cec402-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.733775 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dd66a5e0-0214-485f-80dd-0a54e4cec402-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.733788 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd66a5e0-0214-485f-80dd-0a54e4cec402-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.733800 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgfbr\" (UniqueName: \"kubernetes.io/projected/496c7f69-dbf1-4680-94db-b58b1a0044f2-kube-api-access-mgfbr\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.733813 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496c7f69-dbf1-4680-94db-b58b1a0044f2-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.733826 4934 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/496c7f69-dbf1-4680-94db-b58b1a0044f2-client-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.733838 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2r8s\" (UniqueName: \"kubernetes.io/projected/dd66a5e0-0214-485f-80dd-0a54e4cec402-kube-api-access-s2r8s\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.733878 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496c7f69-dbf1-4680-94db-b58b1a0044f2-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.734605 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f89e5ac0-bc03-4af9-91bb-918412fa9b47-config\") pod \"route-controller-manager-58564d585f-nmxfj\" (UID: \"f89e5ac0-bc03-4af9-91bb-918412fa9b47\") " pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.735019 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f89e5ac0-bc03-4af9-91bb-918412fa9b47-client-ca\") pod \"route-controller-manager-58564d585f-nmxfj\" (UID: \"f89e5ac0-bc03-4af9-91bb-918412fa9b47\") " pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.738188 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f89e5ac0-bc03-4af9-91bb-918412fa9b47-serving-cert\") pod \"route-controller-manager-58564d585f-nmxfj\" (UID: \"f89e5ac0-bc03-4af9-91bb-918412fa9b47\") " pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.754173 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72bd5\" (UniqueName: \"kubernetes.io/projected/f89e5ac0-bc03-4af9-91bb-918412fa9b47-kube-api-access-72bd5\") pod \"route-controller-manager-58564d585f-nmxfj\" (UID: \"f89e5ac0-bc03-4af9-91bb-918412fa9b47\") " pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" Dec 27 07:46:17 crc kubenswrapper[4934]: I1227 07:46:17.890609 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.176127 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" event={"ID":"496c7f69-dbf1-4680-94db-b58b1a0044f2","Type":"ContainerDied","Data":"60cee34f2d024194401e12274e4a5f74c701af376edcd86067713ae94100ec24"} Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.176185 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm" Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.176213 4934 scope.go:117] "RemoveContainer" containerID="79a560e21aded6e113b938f5956cda93ad5a2949534d477f0d1ba72d62f91631" Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.178824 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.178810 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-665c7b5866-brtpg" event={"ID":"dd66a5e0-0214-485f-80dd-0a54e4cec402","Type":"ContainerDied","Data":"1f1519149193dff3409e9f7ef4b290b4238ee1fd2d2e6a4875ca0daf0bae6a55"} Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.219320 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-665c7b5866-brtpg"] Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.226143 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-665c7b5866-brtpg"] Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.232235 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm"] Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.236382 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-796dfb4895-65ztm"] Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.587076 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gk922" Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.598741 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gtrfs" Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.748302 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c26588b0-e09e-4bac-bfd4-22289100341e-utilities\") pod \"c26588b0-e09e-4bac-bfd4-22289100341e\" (UID: \"c26588b0-e09e-4bac-bfd4-22289100341e\") " Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.748733 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f3aa628-fe89-4de2-ad21-34b7d18c9033-catalog-content\") pod \"5f3aa628-fe89-4de2-ad21-34b7d18c9033\" (UID: \"5f3aa628-fe89-4de2-ad21-34b7d18c9033\") " Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.748908 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c26588b0-e09e-4bac-bfd4-22289100341e-catalog-content\") pod \"c26588b0-e09e-4bac-bfd4-22289100341e\" (UID: \"c26588b0-e09e-4bac-bfd4-22289100341e\") " Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.749189 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f3aa628-fe89-4de2-ad21-34b7d18c9033-utilities\") pod \"5f3aa628-fe89-4de2-ad21-34b7d18c9033\" (UID: \"5f3aa628-fe89-4de2-ad21-34b7d18c9033\") " Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.749417 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6544m\" (UniqueName: \"kubernetes.io/projected/c26588b0-e09e-4bac-bfd4-22289100341e-kube-api-access-6544m\") pod \"c26588b0-e09e-4bac-bfd4-22289100341e\" (UID: \"c26588b0-e09e-4bac-bfd4-22289100341e\") " Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.750174 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvtqw\" (UniqueName: \"kubernetes.io/projected/5f3aa628-fe89-4de2-ad21-34b7d18c9033-kube-api-access-vvtqw\") pod \"5f3aa628-fe89-4de2-ad21-34b7d18c9033\" (UID: \"5f3aa628-fe89-4de2-ad21-34b7d18c9033\") " Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.750191 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f3aa628-fe89-4de2-ad21-34b7d18c9033-utilities" (OuterVolumeSpecName: "utilities") pod "5f3aa628-fe89-4de2-ad21-34b7d18c9033" (UID: "5f3aa628-fe89-4de2-ad21-34b7d18c9033"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.749238 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c26588b0-e09e-4bac-bfd4-22289100341e-utilities" (OuterVolumeSpecName: "utilities") pod "c26588b0-e09e-4bac-bfd4-22289100341e" (UID: "c26588b0-e09e-4bac-bfd4-22289100341e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.750993 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c26588b0-e09e-4bac-bfd4-22289100341e-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.751261 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f3aa628-fe89-4de2-ad21-34b7d18c9033-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.754204 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c26588b0-e09e-4bac-bfd4-22289100341e-kube-api-access-6544m" (OuterVolumeSpecName: "kube-api-access-6544m") pod "c26588b0-e09e-4bac-bfd4-22289100341e" (UID: "c26588b0-e09e-4bac-bfd4-22289100341e"). InnerVolumeSpecName "kube-api-access-6544m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.754504 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f3aa628-fe89-4de2-ad21-34b7d18c9033-kube-api-access-vvtqw" (OuterVolumeSpecName: "kube-api-access-vvtqw") pod "5f3aa628-fe89-4de2-ad21-34b7d18c9033" (UID: "5f3aa628-fe89-4de2-ad21-34b7d18c9033"). InnerVolumeSpecName "kube-api-access-vvtqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.827275 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f3aa628-fe89-4de2-ad21-34b7d18c9033-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f3aa628-fe89-4de2-ad21-34b7d18c9033" (UID: "5f3aa628-fe89-4de2-ad21-34b7d18c9033"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.852415 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f3aa628-fe89-4de2-ad21-34b7d18c9033-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.852592 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6544m\" (UniqueName: \"kubernetes.io/projected/c26588b0-e09e-4bac-bfd4-22289100341e-kube-api-access-6544m\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:18 crc kubenswrapper[4934]: I1227 07:46:18.852712 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvtqw\" (UniqueName: \"kubernetes.io/projected/5f3aa628-fe89-4de2-ad21-34b7d18c9033-kube-api-access-vvtqw\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.097674 4934 scope.go:117] "RemoveContainer" containerID="806d607c93d8584bd0c4554a5b4d3ddf7e88a2f57ae29a6b60025905f4d866f4" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.188993 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gk922" event={"ID":"5f3aa628-fe89-4de2-ad21-34b7d18c9033","Type":"ContainerDied","Data":"3b80adbdf4ad62e2df8313f61e990f658f1c3539e536bee70b1ecf9ccd6d3580"} Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.189111 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gk922" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.191997 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gtrfs" event={"ID":"c26588b0-e09e-4bac-bfd4-22289100341e","Type":"ContainerDied","Data":"0728636b829e3e65d843ed46e682384caf901f66408a34a8522dd3473ac3c8ed"} Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.192649 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gtrfs" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.194373 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lf8vn" podUID="4d8b26f6-cfe5-44d8-89e2-92fff18c71a7" containerName="registry-server" containerID="cri-o://4ce2add7f5e513f5a9c3021049bc64929393929481adb6ccfe7765397a5fa8a4" gracePeriod=2 Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.230164 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gk922"] Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.234804 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gk922"] Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.339858 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c26588b0-e09e-4bac-bfd4-22289100341e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c26588b0-e09e-4bac-bfd4-22289100341e" (UID: "c26588b0-e09e-4bac-bfd4-22289100341e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.360123 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c26588b0-e09e-4bac-bfd4-22289100341e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.475737 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496c7f69-dbf1-4680-94db-b58b1a0044f2" path="/var/lib/kubelet/pods/496c7f69-dbf1-4680-94db-b58b1a0044f2/volumes" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.476214 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f3aa628-fe89-4de2-ad21-34b7d18c9033" path="/var/lib/kubelet/pods/5f3aa628-fe89-4de2-ad21-34b7d18c9033/volumes" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.476821 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd66a5e0-0214-485f-80dd-0a54e4cec402" path="/var/lib/kubelet/pods/dd66a5e0-0214-485f-80dd-0a54e4cec402/volumes" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.507607 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gtrfs"] Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.511353 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gtrfs"] Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.612166 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w"] Dec 27 07:46:19 crc kubenswrapper[4934]: E1227 07:46:19.612442 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c26588b0-e09e-4bac-bfd4-22289100341e" containerName="registry-server" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.612462 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c26588b0-e09e-4bac-bfd4-22289100341e" containerName="registry-server" Dec 27 07:46:19 crc kubenswrapper[4934]: E1227 07:46:19.612482 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f3aa628-fe89-4de2-ad21-34b7d18c9033" containerName="extract-utilities" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.612493 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f3aa628-fe89-4de2-ad21-34b7d18c9033" containerName="extract-utilities" Dec 27 07:46:19 crc kubenswrapper[4934]: E1227 07:46:19.612513 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c26588b0-e09e-4bac-bfd4-22289100341e" containerName="extract-content" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.612522 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c26588b0-e09e-4bac-bfd4-22289100341e" containerName="extract-content" Dec 27 07:46:19 crc kubenswrapper[4934]: E1227 07:46:19.612536 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f3aa628-fe89-4de2-ad21-34b7d18c9033" containerName="extract-content" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.612545 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f3aa628-fe89-4de2-ad21-34b7d18c9033" containerName="extract-content" Dec 27 07:46:19 crc kubenswrapper[4934]: E1227 07:46:19.612560 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f3aa628-fe89-4de2-ad21-34b7d18c9033" containerName="registry-server" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.612568 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f3aa628-fe89-4de2-ad21-34b7d18c9033" containerName="registry-server" Dec 27 07:46:19 crc kubenswrapper[4934]: E1227 07:46:19.612580 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c26588b0-e09e-4bac-bfd4-22289100341e" containerName="extract-utilities" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.612589 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c26588b0-e09e-4bac-bfd4-22289100341e" containerName="extract-utilities" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.612696 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c26588b0-e09e-4bac-bfd4-22289100341e" containerName="registry-server" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.612711 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f3aa628-fe89-4de2-ad21-34b7d18c9033" containerName="registry-server" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.613217 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.616682 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.623721 4934 scope.go:117] "RemoveContainer" containerID="09dbaed5045f6892fb85182c594fb8fdb229669203a70d9713f40e7f7afe6dca" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.623784 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.623982 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.624097 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.624135 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w"] Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.624239 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.625444 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.628360 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.731423 4934 scope.go:117] "RemoveContainer" containerID="4b1029c1f10283da6f1ef020f2ae506a5768e782b079e1e0accbb92ed4c3c21d" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.766052 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32619f39-2bc4-4208-bd1d-9d715cf2ad09-serving-cert\") pod \"controller-manager-5d6b58cf7d-hgq5w\" (UID: \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\") " pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.766380 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32619f39-2bc4-4208-bd1d-9d715cf2ad09-client-ca\") pod \"controller-manager-5d6b58cf7d-hgq5w\" (UID: \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\") " pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.766513 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtndr\" (UniqueName: \"kubernetes.io/projected/32619f39-2bc4-4208-bd1d-9d715cf2ad09-kube-api-access-gtndr\") pod \"controller-manager-5d6b58cf7d-hgq5w\" (UID: \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\") " pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.766618 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32619f39-2bc4-4208-bd1d-9d715cf2ad09-config\") pod \"controller-manager-5d6b58cf7d-hgq5w\" (UID: \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\") " pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.766729 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/32619f39-2bc4-4208-bd1d-9d715cf2ad09-proxy-ca-bundles\") pod \"controller-manager-5d6b58cf7d-hgq5w\" (UID: \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\") " pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.788328 4934 scope.go:117] "RemoveContainer" containerID="a2c2c105ffe56da3716b3298c27e762726c6656192045ae2ebd324cba35b7b36" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.832355 4934 scope.go:117] "RemoveContainer" containerID="c9cfec9cae89d6648484ad3b175a8774da0551baa5946d59da5555273e9fac61" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.853111 4934 scope.go:117] "RemoveContainer" containerID="ef3d8fcc3134c47c91d6bf06abe176a1d3915e3bd245d03b5e91b31d9082697e" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.865115 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj"] Dec 27 07:46:19 crc kubenswrapper[4934]: W1227 07:46:19.866798 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf89e5ac0_bc03_4af9_91bb_918412fa9b47.slice/crio-fe76c6bc684aa45ccbbcf5c35202afc840d77c96fc096bd6c159532242204bff WatchSource:0}: Error finding container fe76c6bc684aa45ccbbcf5c35202afc840d77c96fc096bd6c159532242204bff: Status 404 returned error can't find the container with id fe76c6bc684aa45ccbbcf5c35202afc840d77c96fc096bd6c159532242204bff Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.867411 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32619f39-2bc4-4208-bd1d-9d715cf2ad09-client-ca\") pod \"controller-manager-5d6b58cf7d-hgq5w\" (UID: \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\") " pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.867455 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtndr\" (UniqueName: \"kubernetes.io/projected/32619f39-2bc4-4208-bd1d-9d715cf2ad09-kube-api-access-gtndr\") pod \"controller-manager-5d6b58cf7d-hgq5w\" (UID: \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\") " pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.867487 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32619f39-2bc4-4208-bd1d-9d715cf2ad09-config\") pod \"controller-manager-5d6b58cf7d-hgq5w\" (UID: \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\") " pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.867505 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/32619f39-2bc4-4208-bd1d-9d715cf2ad09-proxy-ca-bundles\") pod \"controller-manager-5d6b58cf7d-hgq5w\" (UID: \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\") " pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.867529 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32619f39-2bc4-4208-bd1d-9d715cf2ad09-serving-cert\") pod \"controller-manager-5d6b58cf7d-hgq5w\" (UID: \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\") " pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.868616 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/32619f39-2bc4-4208-bd1d-9d715cf2ad09-proxy-ca-bundles\") pod \"controller-manager-5d6b58cf7d-hgq5w\" (UID: \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\") " pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.868767 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32619f39-2bc4-4208-bd1d-9d715cf2ad09-client-ca\") pod \"controller-manager-5d6b58cf7d-hgq5w\" (UID: \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\") " pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.869877 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32619f39-2bc4-4208-bd1d-9d715cf2ad09-config\") pod \"controller-manager-5d6b58cf7d-hgq5w\" (UID: \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\") " pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.874607 4934 scope.go:117] "RemoveContainer" containerID="f6733cebfbca029ad99eaa4727ce1408064affd12a8ccd6008d9422faad893cf" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.875528 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32619f39-2bc4-4208-bd1d-9d715cf2ad09-serving-cert\") pod \"controller-manager-5d6b58cf7d-hgq5w\" (UID: \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\") " pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" Dec 27 07:46:19 crc kubenswrapper[4934]: I1227 07:46:19.885290 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtndr\" (UniqueName: \"kubernetes.io/projected/32619f39-2bc4-4208-bd1d-9d715cf2ad09-kube-api-access-gtndr\") pod \"controller-manager-5d6b58cf7d-hgq5w\" (UID: \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\") " pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" Dec 27 07:46:20 crc kubenswrapper[4934]: I1227 07:46:20.129380 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" Dec 27 07:46:20 crc kubenswrapper[4934]: I1227 07:46:20.210730 4934 generic.go:334] "Generic (PLEG): container finished" podID="4d8b26f6-cfe5-44d8-89e2-92fff18c71a7" containerID="4ce2add7f5e513f5a9c3021049bc64929393929481adb6ccfe7765397a5fa8a4" exitCode=0 Dec 27 07:46:20 crc kubenswrapper[4934]: I1227 07:46:20.210794 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lf8vn" event={"ID":"4d8b26f6-cfe5-44d8-89e2-92fff18c71a7","Type":"ContainerDied","Data":"4ce2add7f5e513f5a9c3021049bc64929393929481adb6ccfe7765397a5fa8a4"} Dec 27 07:46:20 crc kubenswrapper[4934]: I1227 07:46:20.212263 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" event={"ID":"f89e5ac0-bc03-4af9-91bb-918412fa9b47","Type":"ContainerStarted","Data":"fe76c6bc684aa45ccbbcf5c35202afc840d77c96fc096bd6c159532242204bff"} Dec 27 07:46:20 crc kubenswrapper[4934]: I1227 07:46:20.560699 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w"] Dec 27 07:46:20 crc kubenswrapper[4934]: W1227 07:46:20.565812 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32619f39_2bc4_4208_bd1d_9d715cf2ad09.slice/crio-b67a8ba7d01ca1cc88079b87f884cbac1b8fe5de1f1892cde8216eab64d8009c WatchSource:0}: Error finding container b67a8ba7d01ca1cc88079b87f884cbac1b8fe5de1f1892cde8216eab64d8009c: Status 404 returned error can't find the container with id b67a8ba7d01ca1cc88079b87f884cbac1b8fe5de1f1892cde8216eab64d8009c Dec 27 07:46:20 crc kubenswrapper[4934]: I1227 07:46:20.614736 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lf8vn" Dec 27 07:46:20 crc kubenswrapper[4934]: I1227 07:46:20.779435 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d8b26f6-cfe5-44d8-89e2-92fff18c71a7-catalog-content\") pod \"4d8b26f6-cfe5-44d8-89e2-92fff18c71a7\" (UID: \"4d8b26f6-cfe5-44d8-89e2-92fff18c71a7\") " Dec 27 07:46:20 crc kubenswrapper[4934]: I1227 07:46:20.779535 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d8b26f6-cfe5-44d8-89e2-92fff18c71a7-utilities\") pod \"4d8b26f6-cfe5-44d8-89e2-92fff18c71a7\" (UID: \"4d8b26f6-cfe5-44d8-89e2-92fff18c71a7\") " Dec 27 07:46:20 crc kubenswrapper[4934]: I1227 07:46:20.779613 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdz7v\" (UniqueName: \"kubernetes.io/projected/4d8b26f6-cfe5-44d8-89e2-92fff18c71a7-kube-api-access-zdz7v\") pod \"4d8b26f6-cfe5-44d8-89e2-92fff18c71a7\" (UID: \"4d8b26f6-cfe5-44d8-89e2-92fff18c71a7\") " Dec 27 07:46:20 crc kubenswrapper[4934]: I1227 07:46:20.780959 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d8b26f6-cfe5-44d8-89e2-92fff18c71a7-utilities" (OuterVolumeSpecName: "utilities") pod "4d8b26f6-cfe5-44d8-89e2-92fff18c71a7" (UID: "4d8b26f6-cfe5-44d8-89e2-92fff18c71a7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:46:20 crc kubenswrapper[4934]: I1227 07:46:20.785647 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d8b26f6-cfe5-44d8-89e2-92fff18c71a7-kube-api-access-zdz7v" (OuterVolumeSpecName: "kube-api-access-zdz7v") pod "4d8b26f6-cfe5-44d8-89e2-92fff18c71a7" (UID: "4d8b26f6-cfe5-44d8-89e2-92fff18c71a7"). InnerVolumeSpecName "kube-api-access-zdz7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:46:20 crc kubenswrapper[4934]: I1227 07:46:20.881219 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d8b26f6-cfe5-44d8-89e2-92fff18c71a7-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:20 crc kubenswrapper[4934]: I1227 07:46:20.881256 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdz7v\" (UniqueName: \"kubernetes.io/projected/4d8b26f6-cfe5-44d8-89e2-92fff18c71a7-kube-api-access-zdz7v\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:20 crc kubenswrapper[4934]: I1227 07:46:20.902006 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d8b26f6-cfe5-44d8-89e2-92fff18c71a7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d8b26f6-cfe5-44d8-89e2-92fff18c71a7" (UID: "4d8b26f6-cfe5-44d8-89e2-92fff18c71a7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:46:20 crc kubenswrapper[4934]: I1227 07:46:20.982125 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d8b26f6-cfe5-44d8-89e2-92fff18c71a7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.222542 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" event={"ID":"f89e5ac0-bc03-4af9-91bb-918412fa9b47","Type":"ContainerStarted","Data":"fa149a7dc9e65f9245738329973d99cf95574f29899cdb8b9bfbe54a63964440"} Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.222904 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.226330 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5phzx" event={"ID":"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20","Type":"ContainerStarted","Data":"8fc094e76b88a30ad3289a8824be3a1ccdc50ac6b968a5d4dc4084c14dfe9a3f"} Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.228287 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lzcth" event={"ID":"52a38d21-a579-4f37-be2f-856e2f837f83","Type":"ContainerStarted","Data":"26bbc552e39fe2ce308bd7dbb37a1c790c3983485fa6de88109a88cb0aab90b1"} Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.229636 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" event={"ID":"32619f39-2bc4-4208-bd1d-9d715cf2ad09","Type":"ContainerStarted","Data":"1b261a8503d149ca4b20616f4a66b4714f8e7c85bd4380fd3a6570572fd8cf41"} Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.229664 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" event={"ID":"32619f39-2bc4-4208-bd1d-9d715cf2ad09","Type":"ContainerStarted","Data":"b67a8ba7d01ca1cc88079b87f884cbac1b8fe5de1f1892cde8216eab64d8009c"} Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.230067 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.230593 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.231663 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lf8vn" event={"ID":"4d8b26f6-cfe5-44d8-89e2-92fff18c71a7","Type":"ContainerDied","Data":"18251686229c251ae27fa8528e64f334beca0386ecb27f257873591315f10162"} Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.231707 4934 scope.go:117] "RemoveContainer" containerID="4ce2add7f5e513f5a9c3021049bc64929393929481adb6ccfe7765397a5fa8a4" Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.231738 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lf8vn" Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.234599 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.245469 4934 scope.go:117] "RemoveContainer" containerID="d3b81630a2af7e42360f9a05fd67fd96cc75605a7b5e7cce2fed372b0504bd27" Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.247179 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" podStartSLOduration=12.247167425 podStartE2EDuration="12.247167425s" podCreationTimestamp="2025-12-27 07:46:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:46:21.245046363 +0000 UTC m=+242.065486967" watchObservedRunningTime="2025-12-27 07:46:21.247167425 +0000 UTC m=+242.067608019" Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.271193 4934 scope.go:117] "RemoveContainer" containerID="b38f9f3b63ec18284f1b6d6f2c12cf4da720f9cde02731edab61293749bce9ee" Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.281499 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lzcth" podStartSLOduration=5.257181307 podStartE2EDuration="1m29.281478967s" podCreationTimestamp="2025-12-27 07:44:52 +0000 UTC" firstStartedPulling="2025-12-27 07:44:55.570229085 +0000 UTC m=+156.390669679" lastFinishedPulling="2025-12-27 07:46:19.594526745 +0000 UTC m=+240.414967339" observedRunningTime="2025-12-27 07:46:21.276604303 +0000 UTC m=+242.097044907" watchObservedRunningTime="2025-12-27 07:46:21.281478967 +0000 UTC m=+242.101919551" Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.380628 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" podStartSLOduration=12.380611201 podStartE2EDuration="12.380611201s" podCreationTimestamp="2025-12-27 07:46:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:46:21.379342263 +0000 UTC m=+242.199782857" watchObservedRunningTime="2025-12-27 07:46:21.380611201 +0000 UTC m=+242.201051795" Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.382298 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5phzx" podStartSLOduration=5.174666869 podStartE2EDuration="1m31.38229137s" podCreationTimestamp="2025-12-27 07:44:50 +0000 UTC" firstStartedPulling="2025-12-27 07:44:52.89003399 +0000 UTC m=+153.710474574" lastFinishedPulling="2025-12-27 07:46:19.097658471 +0000 UTC m=+239.918099075" observedRunningTime="2025-12-27 07:46:21.361416695 +0000 UTC m=+242.181857289" watchObservedRunningTime="2025-12-27 07:46:21.38229137 +0000 UTC m=+242.202731964" Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.390361 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lf8vn"] Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.395376 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lf8vn"] Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.474289 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d8b26f6-cfe5-44d8-89e2-92fff18c71a7" path="/var/lib/kubelet/pods/4d8b26f6-cfe5-44d8-89e2-92fff18c71a7/volumes" Dec 27 07:46:21 crc kubenswrapper[4934]: I1227 07:46:21.474885 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c26588b0-e09e-4bac-bfd4-22289100341e" path="/var/lib/kubelet/pods/c26588b0-e09e-4bac-bfd4-22289100341e/volumes" Dec 27 07:46:23 crc kubenswrapper[4934]: I1227 07:46:23.481571 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lzcth" Dec 27 07:46:23 crc kubenswrapper[4934]: I1227 07:46:23.481908 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lzcth" Dec 27 07:46:23 crc kubenswrapper[4934]: I1227 07:46:23.509458 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lzcth" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.741840 4934 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 27 07:46:26 crc kubenswrapper[4934]: E1227 07:46:26.742291 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d8b26f6-cfe5-44d8-89e2-92fff18c71a7" containerName="registry-server" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.742302 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d8b26f6-cfe5-44d8-89e2-92fff18c71a7" containerName="registry-server" Dec 27 07:46:26 crc kubenswrapper[4934]: E1227 07:46:26.742317 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d8b26f6-cfe5-44d8-89e2-92fff18c71a7" containerName="extract-content" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.742322 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d8b26f6-cfe5-44d8-89e2-92fff18c71a7" containerName="extract-content" Dec 27 07:46:26 crc kubenswrapper[4934]: E1227 07:46:26.742336 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d8b26f6-cfe5-44d8-89e2-92fff18c71a7" containerName="extract-utilities" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.742343 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d8b26f6-cfe5-44d8-89e2-92fff18c71a7" containerName="extract-utilities" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.742468 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d8b26f6-cfe5-44d8-89e2-92fff18c71a7" containerName="registry-server" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.742810 4934 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.742968 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.743039 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787" gracePeriod=15 Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.743104 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a" gracePeriod=15 Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.743270 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4" gracePeriod=15 Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.743253 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9" gracePeriod=15 Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.744449 4934 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 27 07:46:26 crc kubenswrapper[4934]: E1227 07:46:26.744902 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.744918 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 27 07:46:26 crc kubenswrapper[4934]: E1227 07:46:26.744933 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.744941 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 27 07:46:26 crc kubenswrapper[4934]: E1227 07:46:26.744960 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.744968 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 27 07:46:26 crc kubenswrapper[4934]: E1227 07:46:26.744978 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.744985 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 27 07:46:26 crc kubenswrapper[4934]: E1227 07:46:26.744995 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.745003 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 27 07:46:26 crc kubenswrapper[4934]: E1227 07:46:26.745014 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.745022 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.745210 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.745227 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.745237 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.745250 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.745261 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.743616 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8" gracePeriod=15 Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.779607 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.875655 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.875708 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.875731 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.875749 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.875805 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.875851 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.875993 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.876049 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.978094 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.978155 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.978161 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.978212 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.978175 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.978214 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.978253 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.978272 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.978341 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.978377 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.978394 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.978376 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.978439 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.978525 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.978588 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 27 07:46:26 crc kubenswrapper[4934]: I1227 07:46:26.978608 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:46:27 crc kubenswrapper[4934]: I1227 07:46:27.076696 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 27 07:46:27 crc kubenswrapper[4934]: W1227 07:46:27.117788 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-b35b60ed2c6cb29491efa4bfe9947098e84e62ec66f921a30933b86bf71df8e4 WatchSource:0}: Error finding container b35b60ed2c6cb29491efa4bfe9947098e84e62ec66f921a30933b86bf71df8e4: Status 404 returned error can't find the container with id b35b60ed2c6cb29491efa4bfe9947098e84e62ec66f921a30933b86bf71df8e4 Dec 27 07:46:27 crc kubenswrapper[4934]: E1227 07:46:27.138787 4934 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.13:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188502df5acbec1e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-27 07:46:27.137063966 +0000 UTC m=+247.957504560,LastTimestamp:2025-12-27 07:46:27.137063966 +0000 UTC m=+247.957504560,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 27 07:46:27 crc kubenswrapper[4934]: I1227 07:46:27.144180 4934 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Dec 27 07:46:27 crc kubenswrapper[4934]: I1227 07:46:27.144249 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Dec 27 07:46:27 crc kubenswrapper[4934]: I1227 07:46:27.267042 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b35b60ed2c6cb29491efa4bfe9947098e84e62ec66f921a30933b86bf71df8e4"} Dec 27 07:46:27 crc kubenswrapper[4934]: E1227 07:46:27.765440 4934 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.13:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188502df5acbec1e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-27 07:46:27.137063966 +0000 UTC m=+247.957504560,LastTimestamp:2025-12-27 07:46:27.137063966 +0000 UTC m=+247.957504560,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 27 07:46:28 crc kubenswrapper[4934]: I1227 07:46:28.278830 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 27 07:46:28 crc kubenswrapper[4934]: I1227 07:46:28.280244 4934 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9" exitCode=2 Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.289719 4934 generic.go:334] "Generic (PLEG): container finished" podID="b8ce8d56-fde4-44c6-a3df-53d4dbca22d4" containerID="1321590a3c6e68643a10b3d4b7e2a2ec45de9e3034e965e3b58c609d93c3d3a3" exitCode=0 Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.289768 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b8ce8d56-fde4-44c6-a3df-53d4dbca22d4","Type":"ContainerDied","Data":"1321590a3c6e68643a10b3d4b7e2a2ec45de9e3034e965e3b58c609d93c3d3a3"} Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.290588 4934 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.290881 4934 status_manager.go:851] "Failed to get status for pod" podUID="b8ce8d56-fde4-44c6-a3df-53d4dbca22d4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.293718 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"a4cbc79fe5abd8973450dc4962819f4db333ffbc34b1d459a9526114bc510ec5"} Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.294408 4934 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.294655 4934 status_manager.go:851] "Failed to get status for pod" podUID="b8ce8d56-fde4-44c6-a3df-53d4dbca22d4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.296341 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.296928 4934 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a" exitCode=0 Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.297028 4934 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4" exitCode=0 Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.297127 4934 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8" exitCode=0 Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.470252 4934 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.470662 4934 status_manager.go:851] "Failed to get status for pod" podUID="b8ce8d56-fde4-44c6-a3df-53d4dbca22d4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.647332 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.647991 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.648683 4934 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.649208 4934 status_manager.go:851] "Failed to get status for pod" podUID="b8ce8d56-fde4-44c6-a3df-53d4dbca22d4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.649691 4934 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.716327 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.716407 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.716434 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.716452 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.716490 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.716522 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.716936 4934 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.716954 4934 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:29 crc kubenswrapper[4934]: I1227 07:46:29.716964 4934 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.308732 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.309814 4934 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787" exitCode=0 Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.309884 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.309953 4934 scope.go:117] "RemoveContainer" containerID="45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.326211 4934 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.326581 4934 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.327277 4934 status_manager.go:851] "Failed to get status for pod" podUID="b8ce8d56-fde4-44c6-a3df-53d4dbca22d4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.340170 4934 scope.go:117] "RemoveContainer" containerID="69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.366433 4934 scope.go:117] "RemoveContainer" containerID="0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.387586 4934 scope.go:117] "RemoveContainer" containerID="b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.409003 4934 scope.go:117] "RemoveContainer" containerID="76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.440258 4934 scope.go:117] "RemoveContainer" containerID="047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.467455 4934 scope.go:117] "RemoveContainer" containerID="45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a" Dec 27 07:46:30 crc kubenswrapper[4934]: E1227 07:46:30.467960 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\": container with ID starting with 45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a not found: ID does not exist" containerID="45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.468008 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a"} err="failed to get container status \"45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\": rpc error: code = NotFound desc = could not find container \"45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a\": container with ID starting with 45b12b974b06d6bc026f9a798571be439cb26e8fc8ad83a8e5598c8c206dd81a not found: ID does not exist" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.468036 4934 scope.go:117] "RemoveContainer" containerID="69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4" Dec 27 07:46:30 crc kubenswrapper[4934]: E1227 07:46:30.468461 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\": container with ID starting with 69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4 not found: ID does not exist" containerID="69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.468492 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4"} err="failed to get container status \"69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\": rpc error: code = NotFound desc = could not find container \"69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4\": container with ID starting with 69ef1193f8beeedf434ba506dedcf156020f1f31837aa328a81bacac7d136ba4 not found: ID does not exist" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.468516 4934 scope.go:117] "RemoveContainer" containerID="0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8" Dec 27 07:46:30 crc kubenswrapper[4934]: E1227 07:46:30.468745 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\": container with ID starting with 0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8 not found: ID does not exist" containerID="0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.468768 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8"} err="failed to get container status \"0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\": rpc error: code = NotFound desc = could not find container \"0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8\": container with ID starting with 0595833e495683574da439202faa79b62684fd12828b098d812bb14dc200e0c8 not found: ID does not exist" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.468784 4934 scope.go:117] "RemoveContainer" containerID="b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9" Dec 27 07:46:30 crc kubenswrapper[4934]: E1227 07:46:30.468930 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\": container with ID starting with b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9 not found: ID does not exist" containerID="b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.468944 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9"} err="failed to get container status \"b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\": rpc error: code = NotFound desc = could not find container \"b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9\": container with ID starting with b67253104294b48ffb764d334a497fa6b974173c4e1ff0269a1972503aafafe9 not found: ID does not exist" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.468955 4934 scope.go:117] "RemoveContainer" containerID="76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787" Dec 27 07:46:30 crc kubenswrapper[4934]: E1227 07:46:30.469175 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\": container with ID starting with 76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787 not found: ID does not exist" containerID="76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.469191 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787"} err="failed to get container status \"76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\": rpc error: code = NotFound desc = could not find container \"76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787\": container with ID starting with 76280e8db6c413c67af7c82c5bb0254b39dfc90ab4c56f21cc25e2bde68c7787 not found: ID does not exist" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.469203 4934 scope.go:117] "RemoveContainer" containerID="047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e" Dec 27 07:46:30 crc kubenswrapper[4934]: E1227 07:46:30.469385 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\": container with ID starting with 047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e not found: ID does not exist" containerID="047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.469408 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e"} err="failed to get container status \"047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\": rpc error: code = NotFound desc = could not find container \"047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e\": container with ID starting with 047b943a0d9925bdad416aed5e7c10170027d718ac27ea7b893e796539dbf71e not found: ID does not exist" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.700545 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.701117 4934 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.701429 4934 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.701619 4934 status_manager.go:851] "Failed to get status for pod" podUID="b8ce8d56-fde4-44c6-a3df-53d4dbca22d4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.801869 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5phzx" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.801945 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5phzx" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.830817 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b8ce8d56-fde4-44c6-a3df-53d4dbca22d4-var-lock\") pod \"b8ce8d56-fde4-44c6-a3df-53d4dbca22d4\" (UID: \"b8ce8d56-fde4-44c6-a3df-53d4dbca22d4\") " Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.830912 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b8ce8d56-fde4-44c6-a3df-53d4dbca22d4-var-lock" (OuterVolumeSpecName: "var-lock") pod "b8ce8d56-fde4-44c6-a3df-53d4dbca22d4" (UID: "b8ce8d56-fde4-44c6-a3df-53d4dbca22d4"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.831113 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b8ce8d56-fde4-44c6-a3df-53d4dbca22d4-kubelet-dir\") pod \"b8ce8d56-fde4-44c6-a3df-53d4dbca22d4\" (UID: \"b8ce8d56-fde4-44c6-a3df-53d4dbca22d4\") " Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.831130 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b8ce8d56-fde4-44c6-a3df-53d4dbca22d4-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b8ce8d56-fde4-44c6-a3df-53d4dbca22d4" (UID: "b8ce8d56-fde4-44c6-a3df-53d4dbca22d4"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.831165 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b8ce8d56-fde4-44c6-a3df-53d4dbca22d4-kube-api-access\") pod \"b8ce8d56-fde4-44c6-a3df-53d4dbca22d4\" (UID: \"b8ce8d56-fde4-44c6-a3df-53d4dbca22d4\") " Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.831383 4934 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b8ce8d56-fde4-44c6-a3df-53d4dbca22d4-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.831398 4934 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b8ce8d56-fde4-44c6-a3df-53d4dbca22d4-var-lock\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.839168 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8ce8d56-fde4-44c6-a3df-53d4dbca22d4-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b8ce8d56-fde4-44c6-a3df-53d4dbca22d4" (UID: "b8ce8d56-fde4-44c6-a3df-53d4dbca22d4"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.855068 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5phzx" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.855499 4934 status_manager.go:851] "Failed to get status for pod" podUID="ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" pod="openshift-marketplace/certified-operators-5phzx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5phzx\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.855927 4934 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.856525 4934 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.856875 4934 status_manager.go:851] "Failed to get status for pod" podUID="b8ce8d56-fde4-44c6-a3df-53d4dbca22d4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:30 crc kubenswrapper[4934]: I1227 07:46:30.932327 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b8ce8d56-fde4-44c6-a3df-53d4dbca22d4-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 27 07:46:31 crc kubenswrapper[4934]: I1227 07:46:31.315482 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"b8ce8d56-fde4-44c6-a3df-53d4dbca22d4","Type":"ContainerDied","Data":"601e4c6339e39ddf20b8f28c462c006536443819240f41e922def1929c2b11f5"} Dec 27 07:46:31 crc kubenswrapper[4934]: I1227 07:46:31.315543 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="601e4c6339e39ddf20b8f28c462c006536443819240f41e922def1929c2b11f5" Dec 27 07:46:31 crc kubenswrapper[4934]: I1227 07:46:31.315498 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 27 07:46:31 crc kubenswrapper[4934]: I1227 07:46:31.337852 4934 status_manager.go:851] "Failed to get status for pod" podUID="ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" pod="openshift-marketplace/certified-operators-5phzx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5phzx\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:31 crc kubenswrapper[4934]: I1227 07:46:31.338632 4934 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:31 crc kubenswrapper[4934]: I1227 07:46:31.339248 4934 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:31 crc kubenswrapper[4934]: I1227 07:46:31.339734 4934 status_manager.go:851] "Failed to get status for pod" podUID="b8ce8d56-fde4-44c6-a3df-53d4dbca22d4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:31 crc kubenswrapper[4934]: I1227 07:46:31.364629 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5phzx" Dec 27 07:46:31 crc kubenswrapper[4934]: I1227 07:46:31.365357 4934 status_manager.go:851] "Failed to get status for pod" podUID="ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" pod="openshift-marketplace/certified-operators-5phzx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5phzx\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:31 crc kubenswrapper[4934]: I1227 07:46:31.365909 4934 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:31 crc kubenswrapper[4934]: I1227 07:46:31.366470 4934 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:31 crc kubenswrapper[4934]: I1227 07:46:31.366926 4934 status_manager.go:851] "Failed to get status for pod" podUID="b8ce8d56-fde4-44c6-a3df-53d4dbca22d4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:31 crc kubenswrapper[4934]: I1227 07:46:31.480579 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 27 07:46:33 crc kubenswrapper[4934]: I1227 07:46:33.522653 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lzcth" Dec 27 07:46:33 crc kubenswrapper[4934]: I1227 07:46:33.523266 4934 status_manager.go:851] "Failed to get status for pod" podUID="ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" pod="openshift-marketplace/certified-operators-5phzx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5phzx\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:33 crc kubenswrapper[4934]: I1227 07:46:33.523588 4934 status_manager.go:851] "Failed to get status for pod" podUID="52a38d21-a579-4f37-be2f-856e2f837f83" pod="openshift-marketplace/redhat-marketplace-lzcth" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-lzcth\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:33 crc kubenswrapper[4934]: I1227 07:46:33.523955 4934 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:33 crc kubenswrapper[4934]: I1227 07:46:33.524344 4934 status_manager.go:851] "Failed to get status for pod" podUID="b8ce8d56-fde4-44c6-a3df-53d4dbca22d4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:34 crc kubenswrapper[4934]: E1227 07:46:34.730069 4934 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:34 crc kubenswrapper[4934]: E1227 07:46:34.730349 4934 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:34 crc kubenswrapper[4934]: E1227 07:46:34.730580 4934 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:34 crc kubenswrapper[4934]: E1227 07:46:34.731156 4934 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:34 crc kubenswrapper[4934]: E1227 07:46:34.731395 4934 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:34 crc kubenswrapper[4934]: I1227 07:46:34.731420 4934 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 27 07:46:34 crc kubenswrapper[4934]: E1227 07:46:34.731634 4934 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" interval="200ms" Dec 27 07:46:34 crc kubenswrapper[4934]: E1227 07:46:34.932470 4934 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" interval="400ms" Dec 27 07:46:35 crc kubenswrapper[4934]: E1227 07:46:35.333707 4934 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" interval="800ms" Dec 27 07:46:36 crc kubenswrapper[4934]: E1227 07:46:36.134306 4934 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" interval="1.6s" Dec 27 07:46:37 crc kubenswrapper[4934]: I1227 07:46:37.466921 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:46:37 crc kubenswrapper[4934]: I1227 07:46:37.467785 4934 status_manager.go:851] "Failed to get status for pod" podUID="ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" pod="openshift-marketplace/certified-operators-5phzx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5phzx\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:37 crc kubenswrapper[4934]: I1227 07:46:37.468015 4934 status_manager.go:851] "Failed to get status for pod" podUID="52a38d21-a579-4f37-be2f-856e2f837f83" pod="openshift-marketplace/redhat-marketplace-lzcth" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-lzcth\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:37 crc kubenswrapper[4934]: I1227 07:46:37.468391 4934 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:37 crc kubenswrapper[4934]: I1227 07:46:37.469070 4934 status_manager.go:851] "Failed to get status for pod" podUID="b8ce8d56-fde4-44c6-a3df-53d4dbca22d4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:37 crc kubenswrapper[4934]: I1227 07:46:37.498731 4934 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="493e2dd1-d571-4aae-8ce2-132231ee766d" Dec 27 07:46:37 crc kubenswrapper[4934]: I1227 07:46:37.499202 4934 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="493e2dd1-d571-4aae-8ce2-132231ee766d" Dec 27 07:46:37 crc kubenswrapper[4934]: E1227 07:46:37.499749 4934 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:46:37 crc kubenswrapper[4934]: I1227 07:46:37.500537 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:46:37 crc kubenswrapper[4934]: W1227 07:46:37.530330 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-5bed919893b4bea42f77666bdf76576c6c526e3b37b4c7ae07f9b3c4a92f6706 WatchSource:0}: Error finding container 5bed919893b4bea42f77666bdf76576c6c526e3b37b4c7ae07f9b3c4a92f6706: Status 404 returned error can't find the container with id 5bed919893b4bea42f77666bdf76576c6c526e3b37b4c7ae07f9b3c4a92f6706 Dec 27 07:46:37 crc kubenswrapper[4934]: E1227 07:46:37.735863 4934 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" interval="3.2s" Dec 27 07:46:37 crc kubenswrapper[4934]: E1227 07:46:37.766356 4934 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.13:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188502df5acbec1e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-27 07:46:27.137063966 +0000 UTC m=+247.957504560,LastTimestamp:2025-12-27 07:46:27.137063966 +0000 UTC m=+247.957504560,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 27 07:46:38 crc kubenswrapper[4934]: I1227 07:46:38.374412 4934 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="b09a9c985c1a4db11ac804b80e50aa54437585d3d32d09d4753bc245a1382ce0" exitCode=0 Dec 27 07:46:38 crc kubenswrapper[4934]: I1227 07:46:38.374510 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"b09a9c985c1a4db11ac804b80e50aa54437585d3d32d09d4753bc245a1382ce0"} Dec 27 07:46:38 crc kubenswrapper[4934]: I1227 07:46:38.374676 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5bed919893b4bea42f77666bdf76576c6c526e3b37b4c7ae07f9b3c4a92f6706"} Dec 27 07:46:38 crc kubenswrapper[4934]: I1227 07:46:38.374926 4934 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="493e2dd1-d571-4aae-8ce2-132231ee766d" Dec 27 07:46:38 crc kubenswrapper[4934]: I1227 07:46:38.374938 4934 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="493e2dd1-d571-4aae-8ce2-132231ee766d" Dec 27 07:46:38 crc kubenswrapper[4934]: E1227 07:46:38.375301 4934 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:46:38 crc kubenswrapper[4934]: I1227 07:46:38.375321 4934 status_manager.go:851] "Failed to get status for pod" podUID="ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" pod="openshift-marketplace/certified-operators-5phzx" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5phzx\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:38 crc kubenswrapper[4934]: I1227 07:46:38.375615 4934 status_manager.go:851] "Failed to get status for pod" podUID="52a38d21-a579-4f37-be2f-856e2f837f83" pod="openshift-marketplace/redhat-marketplace-lzcth" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-lzcth\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:38 crc kubenswrapper[4934]: I1227 07:46:38.375854 4934 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:38 crc kubenswrapper[4934]: I1227 07:46:38.376186 4934 status_manager.go:851] "Failed to get status for pod" podUID="b8ce8d56-fde4-44c6-a3df-53d4dbca22d4" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.13:6443: connect: connection refused" Dec 27 07:46:39 crc kubenswrapper[4934]: I1227 07:46:39.383044 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f4f02975476f6f2aa78b924363d2c6e040a24d57faa1e996417d04d30eb55f51"} Dec 27 07:46:40 crc kubenswrapper[4934]: I1227 07:46:40.395564 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 27 07:46:40 crc kubenswrapper[4934]: I1227 07:46:40.395645 4934 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47" exitCode=1 Dec 27 07:46:40 crc kubenswrapper[4934]: I1227 07:46:40.395712 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47"} Dec 27 07:46:40 crc kubenswrapper[4934]: I1227 07:46:40.396634 4934 scope.go:117] "RemoveContainer" containerID="bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47" Dec 27 07:46:40 crc kubenswrapper[4934]: I1227 07:46:40.401067 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"24c0aafee21a925ea62c14c2fe9fe7431649aa378ad5e822d6bde04804b48dfa"} Dec 27 07:46:42 crc kubenswrapper[4934]: I1227 07:46:42.391920 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:46:42 crc kubenswrapper[4934]: I1227 07:46:42.427831 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 27 07:46:42 crc kubenswrapper[4934]: I1227 07:46:42.428184 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5fa4c8b6a0bef1daea360a8dfb6c71a018b73f127f8fc5c58ef3220b1d5335e7"} Dec 27 07:46:42 crc kubenswrapper[4934]: I1227 07:46:42.431550 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3020cb0a7814d4c1b0f6393b57d83deba6daa82c7e393961ca3fb17e7e358609"} Dec 27 07:46:42 crc kubenswrapper[4934]: I1227 07:46:42.431603 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d488f115f173589bc165adfa860a8d4b8790046f95ad191987de6b4a7c4b72db"} Dec 27 07:46:42 crc kubenswrapper[4934]: I1227 07:46:42.431616 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c272b933d74382eabb2446bf89c03581ca18191a30e8330dc4d45fd1c6f82ea6"} Dec 27 07:46:42 crc kubenswrapper[4934]: I1227 07:46:42.431692 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:46:42 crc kubenswrapper[4934]: I1227 07:46:42.431778 4934 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="493e2dd1-d571-4aae-8ce2-132231ee766d" Dec 27 07:46:42 crc kubenswrapper[4934]: I1227 07:46:42.431798 4934 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="493e2dd1-d571-4aae-8ce2-132231ee766d" Dec 27 07:46:42 crc kubenswrapper[4934]: I1227 07:46:42.439177 4934 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:46:42 crc kubenswrapper[4934]: I1227 07:46:42.501056 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:46:42 crc kubenswrapper[4934]: I1227 07:46:42.501186 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:46:42 crc kubenswrapper[4934]: I1227 07:46:42.506422 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:46:43 crc kubenswrapper[4934]: I1227 07:46:43.438780 4934 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="493e2dd1-d571-4aae-8ce2-132231ee766d" Dec 27 07:46:43 crc kubenswrapper[4934]: I1227 07:46:43.438818 4934 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="493e2dd1-d571-4aae-8ce2-132231ee766d" Dec 27 07:46:43 crc kubenswrapper[4934]: I1227 07:46:43.444123 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:46:44 crc kubenswrapper[4934]: I1227 07:46:44.444615 4934 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="493e2dd1-d571-4aae-8ce2-132231ee766d" Dec 27 07:46:44 crc kubenswrapper[4934]: I1227 07:46:44.445253 4934 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="493e2dd1-d571-4aae-8ce2-132231ee766d" Dec 27 07:46:45 crc kubenswrapper[4934]: I1227 07:46:45.626891 4934 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="16f4603e-b6cb-4f56-a14a-d740f8fa5dc7" Dec 27 07:46:46 crc kubenswrapper[4934]: I1227 07:46:46.353510 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:46:50 crc kubenswrapper[4934]: I1227 07:46:50.046069 4934 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.716880472s: [/var/lib/containers/storage/overlay/b2df7ded895f98b79096b816f672248de3217dc68b3b3c9c9d48daba8d79e81a/diff /var/log/pods/openshift-machine-config-operator_machine-config-daemon-w9j4r_a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7/machine-config-daemon/1.log]; will not log again for this container unless duration exceeds 2s Dec 27 07:46:52 crc kubenswrapper[4934]: I1227 07:46:52.391943 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:46:52 crc kubenswrapper[4934]: I1227 07:46:52.392327 4934 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 27 07:46:52 crc kubenswrapper[4934]: I1227 07:46:52.394589 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 27 07:46:52 crc kubenswrapper[4934]: I1227 07:46:52.976960 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 27 07:46:53 crc kubenswrapper[4934]: I1227 07:46:53.359414 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 27 07:46:53 crc kubenswrapper[4934]: I1227 07:46:53.615335 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 27 07:46:53 crc kubenswrapper[4934]: I1227 07:46:53.843417 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 27 07:46:54 crc kubenswrapper[4934]: I1227 07:46:54.073004 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 27 07:46:54 crc kubenswrapper[4934]: I1227 07:46:54.556417 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 27 07:46:54 crc kubenswrapper[4934]: I1227 07:46:54.573673 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 27 07:46:55 crc kubenswrapper[4934]: I1227 07:46:55.300961 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 27 07:46:56 crc kubenswrapper[4934]: I1227 07:46:56.325932 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 27 07:46:56 crc kubenswrapper[4934]: I1227 07:46:56.441218 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 27 07:46:56 crc kubenswrapper[4934]: I1227 07:46:56.536437 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 27 07:46:56 crc kubenswrapper[4934]: I1227 07:46:56.809457 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 27 07:46:56 crc kubenswrapper[4934]: I1227 07:46:56.987214 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 27 07:46:57 crc kubenswrapper[4934]: I1227 07:46:57.429391 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 27 07:46:57 crc kubenswrapper[4934]: I1227 07:46:57.460117 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 27 07:46:57 crc kubenswrapper[4934]: I1227 07:46:57.638832 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 27 07:46:57 crc kubenswrapper[4934]: I1227 07:46:57.668717 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 27 07:46:57 crc kubenswrapper[4934]: I1227 07:46:57.743779 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 27 07:46:57 crc kubenswrapper[4934]: I1227 07:46:57.851511 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 27 07:46:57 crc kubenswrapper[4934]: I1227 07:46:57.851741 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 27 07:46:57 crc kubenswrapper[4934]: I1227 07:46:57.888465 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 27 07:46:58 crc kubenswrapper[4934]: I1227 07:46:58.058408 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 27 07:46:58 crc kubenswrapper[4934]: I1227 07:46:58.122418 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 27 07:46:58 crc kubenswrapper[4934]: I1227 07:46:58.232967 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 27 07:46:58 crc kubenswrapper[4934]: I1227 07:46:58.318996 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 27 07:46:58 crc kubenswrapper[4934]: I1227 07:46:58.336618 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 27 07:46:58 crc kubenswrapper[4934]: I1227 07:46:58.820457 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 27 07:46:58 crc kubenswrapper[4934]: I1227 07:46:58.969245 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 27 07:46:59 crc kubenswrapper[4934]: I1227 07:46:59.172605 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 27 07:46:59 crc kubenswrapper[4934]: I1227 07:46:59.388046 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 27 07:46:59 crc kubenswrapper[4934]: I1227 07:46:59.698893 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 27 07:46:59 crc kubenswrapper[4934]: I1227 07:46:59.733643 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 27 07:46:59 crc kubenswrapper[4934]: I1227 07:46:59.970234 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 27 07:47:00 crc kubenswrapper[4934]: I1227 07:47:00.042228 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 27 07:47:00 crc kubenswrapper[4934]: I1227 07:47:00.393981 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 27 07:47:00 crc kubenswrapper[4934]: I1227 07:47:00.536653 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 27 07:47:00 crc kubenswrapper[4934]: I1227 07:47:00.550200 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 27 07:47:00 crc kubenswrapper[4934]: I1227 07:47:00.635917 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 27 07:47:00 crc kubenswrapper[4934]: I1227 07:47:00.656813 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 27 07:47:00 crc kubenswrapper[4934]: I1227 07:47:00.835036 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 27 07:47:00 crc kubenswrapper[4934]: I1227 07:47:00.838361 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 27 07:47:00 crc kubenswrapper[4934]: I1227 07:47:00.874396 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 27 07:47:00 crc kubenswrapper[4934]: I1227 07:47:00.892148 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 27 07:47:01 crc kubenswrapper[4934]: I1227 07:47:01.029231 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 27 07:47:01 crc kubenswrapper[4934]: I1227 07:47:01.084857 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 27 07:47:01 crc kubenswrapper[4934]: I1227 07:47:01.262528 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 27 07:47:01 crc kubenswrapper[4934]: I1227 07:47:01.301430 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 27 07:47:01 crc kubenswrapper[4934]: I1227 07:47:01.386841 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 27 07:47:01 crc kubenswrapper[4934]: I1227 07:47:01.500910 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 27 07:47:01 crc kubenswrapper[4934]: I1227 07:47:01.659380 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 27 07:47:01 crc kubenswrapper[4934]: I1227 07:47:01.671391 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 27 07:47:01 crc kubenswrapper[4934]: I1227 07:47:01.736546 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 27 07:47:01 crc kubenswrapper[4934]: I1227 07:47:01.802333 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 27 07:47:01 crc kubenswrapper[4934]: I1227 07:47:01.804789 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 27 07:47:01 crc kubenswrapper[4934]: I1227 07:47:01.809023 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 27 07:47:01 crc kubenswrapper[4934]: I1227 07:47:01.902550 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 27 07:47:01 crc kubenswrapper[4934]: I1227 07:47:01.960665 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 27 07:47:01 crc kubenswrapper[4934]: I1227 07:47:01.981796 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 27 07:47:02 crc kubenswrapper[4934]: I1227 07:47:02.072427 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 27 07:47:02 crc kubenswrapper[4934]: I1227 07:47:02.357959 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 27 07:47:02 crc kubenswrapper[4934]: I1227 07:47:02.392562 4934 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 27 07:47:02 crc kubenswrapper[4934]: I1227 07:47:02.392619 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 27 07:47:02 crc kubenswrapper[4934]: I1227 07:47:02.431818 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 27 07:47:02 crc kubenswrapper[4934]: I1227 07:47:02.505832 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 27 07:47:02 crc kubenswrapper[4934]: I1227 07:47:02.635436 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 27 07:47:02 crc kubenswrapper[4934]: I1227 07:47:02.731044 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 27 07:47:02 crc kubenswrapper[4934]: I1227 07:47:02.741572 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 27 07:47:02 crc kubenswrapper[4934]: I1227 07:47:02.771922 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 27 07:47:02 crc kubenswrapper[4934]: I1227 07:47:02.797162 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 27 07:47:02 crc kubenswrapper[4934]: I1227 07:47:02.854768 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 27 07:47:02 crc kubenswrapper[4934]: I1227 07:47:02.870094 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 27 07:47:02 crc kubenswrapper[4934]: I1227 07:47:02.881202 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 27 07:47:02 crc kubenswrapper[4934]: I1227 07:47:02.931564 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 27 07:47:02 crc kubenswrapper[4934]: I1227 07:47:02.937498 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.081191 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.118862 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.265731 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.370019 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.376302 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.428341 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.436364 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.449477 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.587207 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.672594 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.741235 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.780526 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.842814 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.866909 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.880436 4934 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.883256 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=37.883239446 podStartE2EDuration="37.883239446s" podCreationTimestamp="2025-12-27 07:46:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:46:45.604687306 +0000 UTC m=+266.425127940" watchObservedRunningTime="2025-12-27 07:47:03.883239446 +0000 UTC m=+284.703680040" Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.883446 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.884629 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.884670 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.884687 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lzcth","openshift-marketplace/redhat-marketplace-t4c2w","openshift-marketplace/redhat-operators-xwb5s","openshift-marketplace/marketplace-operator-79b997595-xxrp6","openshift-marketplace/community-operators-wtxbn","openshift-marketplace/certified-operators-5phzx"] Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.884916 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5phzx" podUID="ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" containerName="registry-server" containerID="cri-o://8fc094e76b88a30ad3289a8824be3a1ccdc50ac6b968a5d4dc4084c14dfe9a3f" gracePeriod=30 Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.885757 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" podUID="941a07b0-8984-43f4-8fcd-6cdc8c88f14a" containerName="marketplace-operator" containerID="cri-o://b80c8a2c584e6d6e80879c1af27f4192e3e3344b9730722a768ffc2832867f59" gracePeriod=30 Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.885866 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wtxbn" podUID="b06ee67d-88e4-4c2f-b0a1-efdc7517c40d" containerName="registry-server" containerID="cri-o://0cfd04443627ca28e845e4c8ca2baec3709c700cef244dd8d213200cb5e64c1a" gracePeriod=30 Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.886045 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lzcth" podUID="52a38d21-a579-4f37-be2f-856e2f837f83" containerName="registry-server" containerID="cri-o://26bbc552e39fe2ce308bd7dbb37a1c790c3983485fa6de88109a88cb0aab90b1" gracePeriod=30 Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.886153 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xwb5s" podUID="deaa2d15-bcaf-4568-8847-7bccae9205e0" containerName="registry-server" containerID="cri-o://d0b8d1737acd7fb9e25b1e3efe0bd4d63fd177cc8ee6686b79c81a578ffda4e0" gracePeriod=30 Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.886309 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-t4c2w" podUID="19b07e13-aa5d-4a0a-b6d9-254afc234025" containerName="registry-server" containerID="cri-o://682e5c35be37c30ea824e8afa399bbc5077d2e5832bcf98b79ff6d0c038ecf16" gracePeriod=30 Dec 27 07:47:03 crc kubenswrapper[4934]: I1227 07:47:03.922614 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=21.922600423 podStartE2EDuration="21.922600423s" podCreationTimestamp="2025-12-27 07:46:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:47:03.921336278 +0000 UTC m=+284.741776902" watchObservedRunningTime="2025-12-27 07:47:03.922600423 +0000 UTC m=+284.743041017" Dec 27 07:47:03 crc kubenswrapper[4934]: E1227 07:47:03.951048 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d0b8d1737acd7fb9e25b1e3efe0bd4d63fd177cc8ee6686b79c81a578ffda4e0 is running failed: container process not found" containerID="d0b8d1737acd7fb9e25b1e3efe0bd4d63fd177cc8ee6686b79c81a578ffda4e0" cmd=["grpc_health_probe","-addr=:50051"] Dec 27 07:47:03 crc kubenswrapper[4934]: E1227 07:47:03.951571 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d0b8d1737acd7fb9e25b1e3efe0bd4d63fd177cc8ee6686b79c81a578ffda4e0 is running failed: container process not found" containerID="d0b8d1737acd7fb9e25b1e3efe0bd4d63fd177cc8ee6686b79c81a578ffda4e0" cmd=["grpc_health_probe","-addr=:50051"] Dec 27 07:47:03 crc kubenswrapper[4934]: E1227 07:47:03.952005 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d0b8d1737acd7fb9e25b1e3efe0bd4d63fd177cc8ee6686b79c81a578ffda4e0 is running failed: container process not found" containerID="d0b8d1737acd7fb9e25b1e3efe0bd4d63fd177cc8ee6686b79c81a578ffda4e0" cmd=["grpc_health_probe","-addr=:50051"] Dec 27 07:47:03 crc kubenswrapper[4934]: E1227 07:47:03.952071 4934 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d0b8d1737acd7fb9e25b1e3efe0bd4d63fd177cc8ee6686b79c81a578ffda4e0 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-xwb5s" podUID="deaa2d15-bcaf-4568-8847-7bccae9205e0" containerName="registry-server" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.065492 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.107764 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.147763 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.166779 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.185380 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.216549 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.240384 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.246593 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.308641 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.353099 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.412812 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.467983 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.483206 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.484113 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.516936 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.571602 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.573936 4934 generic.go:334] "Generic (PLEG): container finished" podID="19b07e13-aa5d-4a0a-b6d9-254afc234025" containerID="682e5c35be37c30ea824e8afa399bbc5077d2e5832bcf98b79ff6d0c038ecf16" exitCode=0 Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.573990 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t4c2w" event={"ID":"19b07e13-aa5d-4a0a-b6d9-254afc234025","Type":"ContainerDied","Data":"682e5c35be37c30ea824e8afa399bbc5077d2e5832bcf98b79ff6d0c038ecf16"} Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.576681 4934 generic.go:334] "Generic (PLEG): container finished" podID="52a38d21-a579-4f37-be2f-856e2f837f83" containerID="26bbc552e39fe2ce308bd7dbb37a1c790c3983485fa6de88109a88cb0aab90b1" exitCode=0 Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.576721 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lzcth" event={"ID":"52a38d21-a579-4f37-be2f-856e2f837f83","Type":"ContainerDied","Data":"26bbc552e39fe2ce308bd7dbb37a1c790c3983485fa6de88109a88cb0aab90b1"} Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.578809 4934 generic.go:334] "Generic (PLEG): container finished" podID="b06ee67d-88e4-4c2f-b0a1-efdc7517c40d" containerID="0cfd04443627ca28e845e4c8ca2baec3709c700cef244dd8d213200cb5e64c1a" exitCode=0 Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.578881 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wtxbn" event={"ID":"b06ee67d-88e4-4c2f-b0a1-efdc7517c40d","Type":"ContainerDied","Data":"0cfd04443627ca28e845e4c8ca2baec3709c700cef244dd8d213200cb5e64c1a"} Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.581049 4934 generic.go:334] "Generic (PLEG): container finished" podID="deaa2d15-bcaf-4568-8847-7bccae9205e0" containerID="d0b8d1737acd7fb9e25b1e3efe0bd4d63fd177cc8ee6686b79c81a578ffda4e0" exitCode=0 Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.581101 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwb5s" event={"ID":"deaa2d15-bcaf-4568-8847-7bccae9205e0","Type":"ContainerDied","Data":"d0b8d1737acd7fb9e25b1e3efe0bd4d63fd177cc8ee6686b79c81a578ffda4e0"} Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.582721 4934 generic.go:334] "Generic (PLEG): container finished" podID="941a07b0-8984-43f4-8fcd-6cdc8c88f14a" containerID="b80c8a2c584e6d6e80879c1af27f4192e3e3344b9730722a768ffc2832867f59" exitCode=0 Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.582751 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" event={"ID":"941a07b0-8984-43f4-8fcd-6cdc8c88f14a","Type":"ContainerDied","Data":"b80c8a2c584e6d6e80879c1af27f4192e3e3344b9730722a768ffc2832867f59"} Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.584969 4934 generic.go:334] "Generic (PLEG): container finished" podID="ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" containerID="8fc094e76b88a30ad3289a8824be3a1ccdc50ac6b968a5d4dc4084c14dfe9a3f" exitCode=0 Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.584998 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5phzx" event={"ID":"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20","Type":"ContainerDied","Data":"8fc094e76b88a30ad3289a8824be3a1ccdc50ac6b968a5d4dc4084c14dfe9a3f"} Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.680306 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.728597 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.817732 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.873280 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.877332 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.900317 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t4c2w" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.957175 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.972506 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 27 07:47:04 crc kubenswrapper[4934]: I1227 07:47:04.980757 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.029511 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19b07e13-aa5d-4a0a-b6d9-254afc234025-utilities\") pod \"19b07e13-aa5d-4a0a-b6d9-254afc234025\" (UID: \"19b07e13-aa5d-4a0a-b6d9-254afc234025\") " Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.029660 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19b07e13-aa5d-4a0a-b6d9-254afc234025-catalog-content\") pod \"19b07e13-aa5d-4a0a-b6d9-254afc234025\" (UID: \"19b07e13-aa5d-4a0a-b6d9-254afc234025\") " Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.029705 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2g8n6\" (UniqueName: \"kubernetes.io/projected/19b07e13-aa5d-4a0a-b6d9-254afc234025-kube-api-access-2g8n6\") pod \"19b07e13-aa5d-4a0a-b6d9-254afc234025\" (UID: \"19b07e13-aa5d-4a0a-b6d9-254afc234025\") " Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.031262 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19b07e13-aa5d-4a0a-b6d9-254afc234025-utilities" (OuterVolumeSpecName: "utilities") pod "19b07e13-aa5d-4a0a-b6d9-254afc234025" (UID: "19b07e13-aa5d-4a0a-b6d9-254afc234025"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.034934 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19b07e13-aa5d-4a0a-b6d9-254afc234025-kube-api-access-2g8n6" (OuterVolumeSpecName: "kube-api-access-2g8n6") pod "19b07e13-aa5d-4a0a-b6d9-254afc234025" (UID: "19b07e13-aa5d-4a0a-b6d9-254afc234025"). InnerVolumeSpecName "kube-api-access-2g8n6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.066023 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.066875 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19b07e13-aa5d-4a0a-b6d9-254afc234025-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "19b07e13-aa5d-4a0a-b6d9-254afc234025" (UID: "19b07e13-aa5d-4a0a-b6d9-254afc234025"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.075977 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.079185 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwb5s" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.083908 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wtxbn" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.088320 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5phzx" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.093482 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lzcth" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.131739 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/deaa2d15-bcaf-4568-8847-7bccae9205e0-utilities\") pod \"deaa2d15-bcaf-4568-8847-7bccae9205e0\" (UID: \"deaa2d15-bcaf-4568-8847-7bccae9205e0\") " Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.131792 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20-utilities\") pod \"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20\" (UID: \"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20\") " Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.131831 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gltj\" (UniqueName: \"kubernetes.io/projected/deaa2d15-bcaf-4568-8847-7bccae9205e0-kube-api-access-5gltj\") pod \"deaa2d15-bcaf-4568-8847-7bccae9205e0\" (UID: \"deaa2d15-bcaf-4568-8847-7bccae9205e0\") " Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.131855 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b06ee67d-88e4-4c2f-b0a1-efdc7517c40d-catalog-content\") pod \"b06ee67d-88e4-4c2f-b0a1-efdc7517c40d\" (UID: \"b06ee67d-88e4-4c2f-b0a1-efdc7517c40d\") " Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.131898 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b06ee67d-88e4-4c2f-b0a1-efdc7517c40d-utilities\") pod \"b06ee67d-88e4-4c2f-b0a1-efdc7517c40d\" (UID: \"b06ee67d-88e4-4c2f-b0a1-efdc7517c40d\") " Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.131923 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/941a07b0-8984-43f4-8fcd-6cdc8c88f14a-marketplace-operator-metrics\") pod \"941a07b0-8984-43f4-8fcd-6cdc8c88f14a\" (UID: \"941a07b0-8984-43f4-8fcd-6cdc8c88f14a\") " Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.131944 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftz4h\" (UniqueName: \"kubernetes.io/projected/941a07b0-8984-43f4-8fcd-6cdc8c88f14a-kube-api-access-ftz4h\") pod \"941a07b0-8984-43f4-8fcd-6cdc8c88f14a\" (UID: \"941a07b0-8984-43f4-8fcd-6cdc8c88f14a\") " Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.132006 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/941a07b0-8984-43f4-8fcd-6cdc8c88f14a-marketplace-trusted-ca\") pod \"941a07b0-8984-43f4-8fcd-6cdc8c88f14a\" (UID: \"941a07b0-8984-43f4-8fcd-6cdc8c88f14a\") " Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.132027 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ks49\" (UniqueName: \"kubernetes.io/projected/ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20-kube-api-access-8ks49\") pod \"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20\" (UID: \"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20\") " Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.132050 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xxg7\" (UniqueName: \"kubernetes.io/projected/b06ee67d-88e4-4c2f-b0a1-efdc7517c40d-kube-api-access-6xxg7\") pod \"b06ee67d-88e4-4c2f-b0a1-efdc7517c40d\" (UID: \"b06ee67d-88e4-4c2f-b0a1-efdc7517c40d\") " Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.132077 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qwpc\" (UniqueName: \"kubernetes.io/projected/52a38d21-a579-4f37-be2f-856e2f837f83-kube-api-access-8qwpc\") pod \"52a38d21-a579-4f37-be2f-856e2f837f83\" (UID: \"52a38d21-a579-4f37-be2f-856e2f837f83\") " Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.132166 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52a38d21-a579-4f37-be2f-856e2f837f83-catalog-content\") pod \"52a38d21-a579-4f37-be2f-856e2f837f83\" (UID: \"52a38d21-a579-4f37-be2f-856e2f837f83\") " Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.132235 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20-catalog-content\") pod \"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20\" (UID: \"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20\") " Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.132265 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/deaa2d15-bcaf-4568-8847-7bccae9205e0-catalog-content\") pod \"deaa2d15-bcaf-4568-8847-7bccae9205e0\" (UID: \"deaa2d15-bcaf-4568-8847-7bccae9205e0\") " Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.132294 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52a38d21-a579-4f37-be2f-856e2f837f83-utilities\") pod \"52a38d21-a579-4f37-be2f-856e2f837f83\" (UID: \"52a38d21-a579-4f37-be2f-856e2f837f83\") " Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.132519 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19b07e13-aa5d-4a0a-b6d9-254afc234025-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.132562 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19b07e13-aa5d-4a0a-b6d9-254afc234025-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.132576 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2g8n6\" (UniqueName: \"kubernetes.io/projected/19b07e13-aa5d-4a0a-b6d9-254afc234025-kube-api-access-2g8n6\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.132606 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/deaa2d15-bcaf-4568-8847-7bccae9205e0-utilities" (OuterVolumeSpecName: "utilities") pod "deaa2d15-bcaf-4568-8847-7bccae9205e0" (UID: "deaa2d15-bcaf-4568-8847-7bccae9205e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.133410 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/941a07b0-8984-43f4-8fcd-6cdc8c88f14a-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "941a07b0-8984-43f4-8fcd-6cdc8c88f14a" (UID: "941a07b0-8984-43f4-8fcd-6cdc8c88f14a"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.133485 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52a38d21-a579-4f37-be2f-856e2f837f83-utilities" (OuterVolumeSpecName: "utilities") pod "52a38d21-a579-4f37-be2f-856e2f837f83" (UID: "52a38d21-a579-4f37-be2f-856e2f837f83"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.136049 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b06ee67d-88e4-4c2f-b0a1-efdc7517c40d-utilities" (OuterVolumeSpecName: "utilities") pod "b06ee67d-88e4-4c2f-b0a1-efdc7517c40d" (UID: "b06ee67d-88e4-4c2f-b0a1-efdc7517c40d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.135263 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20-utilities" (OuterVolumeSpecName: "utilities") pod "ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" (UID: "ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.137174 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20-kube-api-access-8ks49" (OuterVolumeSpecName: "kube-api-access-8ks49") pod "ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" (UID: "ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20"). InnerVolumeSpecName "kube-api-access-8ks49". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.138497 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b06ee67d-88e4-4c2f-b0a1-efdc7517c40d-kube-api-access-6xxg7" (OuterVolumeSpecName: "kube-api-access-6xxg7") pod "b06ee67d-88e4-4c2f-b0a1-efdc7517c40d" (UID: "b06ee67d-88e4-4c2f-b0a1-efdc7517c40d"). InnerVolumeSpecName "kube-api-access-6xxg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.140207 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52a38d21-a579-4f37-be2f-856e2f837f83-kube-api-access-8qwpc" (OuterVolumeSpecName: "kube-api-access-8qwpc") pod "52a38d21-a579-4f37-be2f-856e2f837f83" (UID: "52a38d21-a579-4f37-be2f-856e2f837f83"). InnerVolumeSpecName "kube-api-access-8qwpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.140959 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/941a07b0-8984-43f4-8fcd-6cdc8c88f14a-kube-api-access-ftz4h" (OuterVolumeSpecName: "kube-api-access-ftz4h") pod "941a07b0-8984-43f4-8fcd-6cdc8c88f14a" (UID: "941a07b0-8984-43f4-8fcd-6cdc8c88f14a"). InnerVolumeSpecName "kube-api-access-ftz4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.151227 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deaa2d15-bcaf-4568-8847-7bccae9205e0-kube-api-access-5gltj" (OuterVolumeSpecName: "kube-api-access-5gltj") pod "deaa2d15-bcaf-4568-8847-7bccae9205e0" (UID: "deaa2d15-bcaf-4568-8847-7bccae9205e0"). InnerVolumeSpecName "kube-api-access-5gltj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.155892 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/941a07b0-8984-43f4-8fcd-6cdc8c88f14a-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "941a07b0-8984-43f4-8fcd-6cdc8c88f14a" (UID: "941a07b0-8984-43f4-8fcd-6cdc8c88f14a"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.172737 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52a38d21-a579-4f37-be2f-856e2f837f83-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "52a38d21-a579-4f37-be2f-856e2f837f83" (UID: "52a38d21-a579-4f37-be2f-856e2f837f83"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.180237 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.199800 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b06ee67d-88e4-4c2f-b0a1-efdc7517c40d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b06ee67d-88e4-4c2f-b0a1-efdc7517c40d" (UID: "b06ee67d-88e4-4c2f-b0a1-efdc7517c40d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.220661 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" (UID: "ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.233409 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52a38d21-a579-4f37-be2f-856e2f837f83-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.233610 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.233676 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52a38d21-a579-4f37-be2f-856e2f837f83-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.233780 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/deaa2d15-bcaf-4568-8847-7bccae9205e0-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.233842 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.233930 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gltj\" (UniqueName: \"kubernetes.io/projected/deaa2d15-bcaf-4568-8847-7bccae9205e0-kube-api-access-5gltj\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.234010 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b06ee67d-88e4-4c2f-b0a1-efdc7517c40d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.234100 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b06ee67d-88e4-4c2f-b0a1-efdc7517c40d-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.234182 4934 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/941a07b0-8984-43f4-8fcd-6cdc8c88f14a-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.234276 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftz4h\" (UniqueName: \"kubernetes.io/projected/941a07b0-8984-43f4-8fcd-6cdc8c88f14a-kube-api-access-ftz4h\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.234356 4934 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/941a07b0-8984-43f4-8fcd-6cdc8c88f14a-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.234443 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ks49\" (UniqueName: \"kubernetes.io/projected/ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20-kube-api-access-8ks49\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.234519 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xxg7\" (UniqueName: \"kubernetes.io/projected/b06ee67d-88e4-4c2f-b0a1-efdc7517c40d-kube-api-access-6xxg7\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.234588 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qwpc\" (UniqueName: \"kubernetes.io/projected/52a38d21-a579-4f37-be2f-856e2f837f83-kube-api-access-8qwpc\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.254631 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.274977 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/deaa2d15-bcaf-4568-8847-7bccae9205e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "deaa2d15-bcaf-4568-8847-7bccae9205e0" (UID: "deaa2d15-bcaf-4568-8847-7bccae9205e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.335451 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/deaa2d15-bcaf-4568-8847-7bccae9205e0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.392851 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.393634 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.430376 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.490536 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.522704 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.543896 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.544209 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.598901 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" event={"ID":"941a07b0-8984-43f4-8fcd-6cdc8c88f14a","Type":"ContainerDied","Data":"64d5fb0ea26ea5c1ec2c6a50e37de1adb84c0c821483d1514a3979e9bb93a389"} Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.598973 4934 scope.go:117] "RemoveContainer" containerID="b80c8a2c584e6d6e80879c1af27f4192e3e3344b9730722a768ffc2832867f59" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.600019 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xxrp6" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.603525 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5phzx" event={"ID":"ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20","Type":"ContainerDied","Data":"cff48171d21190f88ae52273299a0efe047ac067b630bb9c0c7b6af68d0a42b7"} Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.603662 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5phzx" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.606000 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t4c2w" event={"ID":"19b07e13-aa5d-4a0a-b6d9-254afc234025","Type":"ContainerDied","Data":"8bc7bab559fc2df71fda584267f3d57401360aa3d9305d85d74e16d4345f8855"} Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.606064 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t4c2w" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.609165 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lzcth" event={"ID":"52a38d21-a579-4f37-be2f-856e2f837f83","Type":"ContainerDied","Data":"fdae001289885a13b9cf483490888ea5141f8a90a0fef10f67613acf9f247051"} Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.609283 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lzcth" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.614571 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wtxbn" event={"ID":"b06ee67d-88e4-4c2f-b0a1-efdc7517c40d","Type":"ContainerDied","Data":"ea5f05850cd5912907fe94f65263850a595409b6306d5759b1b7d7cc5426fafc"} Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.614722 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wtxbn" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.618073 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xwb5s" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.618210 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xxrp6"] Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.618256 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xwb5s" event={"ID":"deaa2d15-bcaf-4568-8847-7bccae9205e0","Type":"ContainerDied","Data":"bbc7a319c5e881ac46200c346aeeaea1828f2c079ecedb7ff9d0ac5a7a97bcd6"} Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.623303 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xxrp6"] Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.627224 4934 scope.go:117] "RemoveContainer" containerID="8fc094e76b88a30ad3289a8824be3a1ccdc50ac6b968a5d4dc4084c14dfe9a3f" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.630279 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5phzx"] Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.637922 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5phzx"] Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.643807 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lzcth"] Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.649740 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lzcth"] Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.652391 4934 scope.go:117] "RemoveContainer" containerID="27f3b35dbbee84864846f9a6d9d209cfdb9d2e5bed7d086af8dd62c00a54e841" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.653569 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t4c2w"] Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.661161 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-t4c2w"] Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.667193 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xwb5s"] Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.667897 4934 scope.go:117] "RemoveContainer" containerID="12625e709f755da9b438242ba209f44a8633991928abc9dd8f53f07bdb02d4dd" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.670754 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xwb5s"] Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.673783 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wtxbn"] Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.677453 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wtxbn"] Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.679566 4934 scope.go:117] "RemoveContainer" containerID="682e5c35be37c30ea824e8afa399bbc5077d2e5832bcf98b79ff6d0c038ecf16" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.692454 4934 scope.go:117] "RemoveContainer" containerID="d2c75cb43e3c4cebb027bfd283f92d47af8269e44e26f182db86e42ed276a7d9" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.705112 4934 scope.go:117] "RemoveContainer" containerID="5e32b748d4ec0850d4d6b5aeda4dc212f058207e2b0366c8279afa05c4bd4267" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.715997 4934 scope.go:117] "RemoveContainer" containerID="26bbc552e39fe2ce308bd7dbb37a1c790c3983485fa6de88109a88cb0aab90b1" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.728668 4934 scope.go:117] "RemoveContainer" containerID="8ff79859d5acf8d7d0c0f5eb1aa4bdf590d7a34bcd57d350811631c01fff9670" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.746203 4934 scope.go:117] "RemoveContainer" containerID="597a65e4026ce1b1420029a2105f00a66a748b077e735ba0e448e75d82d38409" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.763368 4934 scope.go:117] "RemoveContainer" containerID="0cfd04443627ca28e845e4c8ca2baec3709c700cef244dd8d213200cb5e64c1a" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.777226 4934 scope.go:117] "RemoveContainer" containerID="71b561737dc74ba67d79f3d9f01f0d37d5d880a14a08a55d05dd373992359eaf" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.791144 4934 scope.go:117] "RemoveContainer" containerID="f7c5cadc3c57d85e2b7791f15387690e5c72a31597f5b0e6470db058f77c0138" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.805503 4934 scope.go:117] "RemoveContainer" containerID="d0b8d1737acd7fb9e25b1e3efe0bd4d63fd177cc8ee6686b79c81a578ffda4e0" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.818128 4934 scope.go:117] "RemoveContainer" containerID="d309456782220cc84313b79fc648c59070da988f5b969378b2b9f9572c4703c8" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.830123 4934 scope.go:117] "RemoveContainer" containerID="fc9f95dec89a8df860380847f53fcda9d618a9c96cd22fe3801f1b278fbe883b" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.831788 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.881054 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 27 07:47:05 crc kubenswrapper[4934]: I1227 07:47:05.968824 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 27 07:47:06 crc kubenswrapper[4934]: I1227 07:47:06.006745 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 27 07:47:06 crc kubenswrapper[4934]: I1227 07:47:06.096528 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 27 07:47:06 crc kubenswrapper[4934]: I1227 07:47:06.243716 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 27 07:47:06 crc kubenswrapper[4934]: I1227 07:47:06.261140 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 27 07:47:06 crc kubenswrapper[4934]: I1227 07:47:06.313017 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 27 07:47:06 crc kubenswrapper[4934]: I1227 07:47:06.461965 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 27 07:47:06 crc kubenswrapper[4934]: I1227 07:47:06.543638 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 27 07:47:06 crc kubenswrapper[4934]: I1227 07:47:06.710514 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 27 07:47:06 crc kubenswrapper[4934]: I1227 07:47:06.716645 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 27 07:47:06 crc kubenswrapper[4934]: I1227 07:47:06.816398 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 27 07:47:06 crc kubenswrapper[4934]: I1227 07:47:06.833982 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 27 07:47:06 crc kubenswrapper[4934]: I1227 07:47:06.849218 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 27 07:47:06 crc kubenswrapper[4934]: I1227 07:47:06.879531 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 27 07:47:06 crc kubenswrapper[4934]: I1227 07:47:06.891505 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 27 07:47:07 crc kubenswrapper[4934]: I1227 07:47:07.127911 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 27 07:47:07 crc kubenswrapper[4934]: I1227 07:47:07.138385 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 27 07:47:07 crc kubenswrapper[4934]: I1227 07:47:07.175785 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 27 07:47:07 crc kubenswrapper[4934]: I1227 07:47:07.190605 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 27 07:47:07 crc kubenswrapper[4934]: I1227 07:47:07.300817 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 27 07:47:07 crc kubenswrapper[4934]: I1227 07:47:07.387679 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 27 07:47:07 crc kubenswrapper[4934]: I1227 07:47:07.406434 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 27 07:47:07 crc kubenswrapper[4934]: I1227 07:47:07.412193 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 27 07:47:07 crc kubenswrapper[4934]: I1227 07:47:07.452591 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 27 07:47:07 crc kubenswrapper[4934]: I1227 07:47:07.475700 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19b07e13-aa5d-4a0a-b6d9-254afc234025" path="/var/lib/kubelet/pods/19b07e13-aa5d-4a0a-b6d9-254afc234025/volumes" Dec 27 07:47:07 crc kubenswrapper[4934]: I1227 07:47:07.476867 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52a38d21-a579-4f37-be2f-856e2f837f83" path="/var/lib/kubelet/pods/52a38d21-a579-4f37-be2f-856e2f837f83/volumes" Dec 27 07:47:07 crc kubenswrapper[4934]: I1227 07:47:07.478112 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="941a07b0-8984-43f4-8fcd-6cdc8c88f14a" path="/var/lib/kubelet/pods/941a07b0-8984-43f4-8fcd-6cdc8c88f14a/volumes" Dec 27 07:47:07 crc kubenswrapper[4934]: I1227 07:47:07.479744 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b06ee67d-88e4-4c2f-b0a1-efdc7517c40d" path="/var/lib/kubelet/pods/b06ee67d-88e4-4c2f-b0a1-efdc7517c40d/volumes" Dec 27 07:47:07 crc kubenswrapper[4934]: I1227 07:47:07.480896 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" path="/var/lib/kubelet/pods/ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20/volumes" Dec 27 07:47:07 crc kubenswrapper[4934]: I1227 07:47:07.483235 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="deaa2d15-bcaf-4568-8847-7bccae9205e0" path="/var/lib/kubelet/pods/deaa2d15-bcaf-4568-8847-7bccae9205e0/volumes" Dec 27 07:47:07 crc kubenswrapper[4934]: I1227 07:47:07.508146 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 27 07:47:07 crc kubenswrapper[4934]: I1227 07:47:07.641340 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 27 07:47:07 crc kubenswrapper[4934]: I1227 07:47:07.682425 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 27 07:47:07 crc kubenswrapper[4934]: I1227 07:47:07.815246 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 27 07:47:07 crc kubenswrapper[4934]: I1227 07:47:07.957481 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 27 07:47:07 crc kubenswrapper[4934]: I1227 07:47:07.982589 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 27 07:47:08 crc kubenswrapper[4934]: I1227 07:47:08.058690 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 27 07:47:08 crc kubenswrapper[4934]: I1227 07:47:08.068732 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 27 07:47:08 crc kubenswrapper[4934]: I1227 07:47:08.120313 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 27 07:47:08 crc kubenswrapper[4934]: I1227 07:47:08.214004 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 27 07:47:08 crc kubenswrapper[4934]: I1227 07:47:08.217652 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 27 07:47:08 crc kubenswrapper[4934]: I1227 07:47:08.232046 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 27 07:47:08 crc kubenswrapper[4934]: I1227 07:47:08.268212 4934 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 27 07:47:08 crc kubenswrapper[4934]: I1227 07:47:08.368037 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 27 07:47:08 crc kubenswrapper[4934]: I1227 07:47:08.376868 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 27 07:47:08 crc kubenswrapper[4934]: I1227 07:47:08.388873 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 27 07:47:08 crc kubenswrapper[4934]: I1227 07:47:08.390799 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 27 07:47:08 crc kubenswrapper[4934]: I1227 07:47:08.442230 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 27 07:47:08 crc kubenswrapper[4934]: I1227 07:47:08.467117 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 27 07:47:08 crc kubenswrapper[4934]: I1227 07:47:08.494749 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 27 07:47:08 crc kubenswrapper[4934]: I1227 07:47:08.505881 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 27 07:47:08 crc kubenswrapper[4934]: I1227 07:47:08.604122 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 27 07:47:08 crc kubenswrapper[4934]: I1227 07:47:08.712645 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 27 07:47:08 crc kubenswrapper[4934]: I1227 07:47:08.872949 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 27 07:47:08 crc kubenswrapper[4934]: I1227 07:47:08.887497 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 27 07:47:08 crc kubenswrapper[4934]: I1227 07:47:08.931331 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 27 07:47:08 crc kubenswrapper[4934]: I1227 07:47:08.979815 4934 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 27 07:47:09 crc kubenswrapper[4934]: I1227 07:47:09.046651 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 27 07:47:09 crc kubenswrapper[4934]: I1227 07:47:09.081053 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 27 07:47:09 crc kubenswrapper[4934]: I1227 07:47:09.121099 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 27 07:47:09 crc kubenswrapper[4934]: I1227 07:47:09.276650 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 27 07:47:09 crc kubenswrapper[4934]: I1227 07:47:09.276704 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 27 07:47:09 crc kubenswrapper[4934]: I1227 07:47:09.527011 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 27 07:47:09 crc kubenswrapper[4934]: I1227 07:47:09.554946 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 27 07:47:09 crc kubenswrapper[4934]: I1227 07:47:09.562689 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 27 07:47:09 crc kubenswrapper[4934]: I1227 07:47:09.666669 4934 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 27 07:47:09 crc kubenswrapper[4934]: I1227 07:47:09.733188 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 27 07:47:09 crc kubenswrapper[4934]: I1227 07:47:09.851546 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 27 07:47:09 crc kubenswrapper[4934]: I1227 07:47:09.866739 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 27 07:47:09 crc kubenswrapper[4934]: I1227 07:47:09.871636 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 27 07:47:09 crc kubenswrapper[4934]: I1227 07:47:09.971486 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 27 07:47:09 crc kubenswrapper[4934]: I1227 07:47:09.989548 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 27 07:47:10 crc kubenswrapper[4934]: I1227 07:47:10.069662 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 27 07:47:10 crc kubenswrapper[4934]: I1227 07:47:10.072701 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 27 07:47:10 crc kubenswrapper[4934]: I1227 07:47:10.105764 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 27 07:47:10 crc kubenswrapper[4934]: I1227 07:47:10.114597 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 27 07:47:10 crc kubenswrapper[4934]: I1227 07:47:10.205458 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 27 07:47:10 crc kubenswrapper[4934]: I1227 07:47:10.256016 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 27 07:47:10 crc kubenswrapper[4934]: I1227 07:47:10.260415 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 27 07:47:10 crc kubenswrapper[4934]: I1227 07:47:10.317744 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 27 07:47:10 crc kubenswrapper[4934]: I1227 07:47:10.365373 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 27 07:47:10 crc kubenswrapper[4934]: I1227 07:47:10.373258 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 27 07:47:10 crc kubenswrapper[4934]: I1227 07:47:10.378236 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 27 07:47:10 crc kubenswrapper[4934]: I1227 07:47:10.494675 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 27 07:47:10 crc kubenswrapper[4934]: I1227 07:47:10.631779 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 27 07:47:10 crc kubenswrapper[4934]: I1227 07:47:10.648621 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 27 07:47:10 crc kubenswrapper[4934]: I1227 07:47:10.863536 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 27 07:47:10 crc kubenswrapper[4934]: I1227 07:47:10.875444 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 27 07:47:10 crc kubenswrapper[4934]: I1227 07:47:10.938635 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 27 07:47:11 crc kubenswrapper[4934]: I1227 07:47:11.064753 4934 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 27 07:47:11 crc kubenswrapper[4934]: I1227 07:47:11.149498 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 27 07:47:11 crc kubenswrapper[4934]: I1227 07:47:11.171982 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 27 07:47:11 crc kubenswrapper[4934]: I1227 07:47:11.253586 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 27 07:47:11 crc kubenswrapper[4934]: I1227 07:47:11.399720 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 27 07:47:11 crc kubenswrapper[4934]: I1227 07:47:11.442988 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 27 07:47:11 crc kubenswrapper[4934]: I1227 07:47:11.516662 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 27 07:47:11 crc kubenswrapper[4934]: I1227 07:47:11.553363 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 27 07:47:11 crc kubenswrapper[4934]: I1227 07:47:11.622762 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 27 07:47:11 crc kubenswrapper[4934]: I1227 07:47:11.637785 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 27 07:47:11 crc kubenswrapper[4934]: I1227 07:47:11.666125 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 27 07:47:11 crc kubenswrapper[4934]: I1227 07:47:11.736826 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 27 07:47:11 crc kubenswrapper[4934]: I1227 07:47:11.893698 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 27 07:47:11 crc kubenswrapper[4934]: I1227 07:47:11.999232 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 27 07:47:12 crc kubenswrapper[4934]: I1227 07:47:12.050571 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 27 07:47:12 crc kubenswrapper[4934]: I1227 07:47:12.391931 4934 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 27 07:47:12 crc kubenswrapper[4934]: I1227 07:47:12.392062 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 27 07:47:12 crc kubenswrapper[4934]: I1227 07:47:12.392155 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:47:12 crc kubenswrapper[4934]: I1227 07:47:12.392863 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"5fa4c8b6a0bef1daea360a8dfb6c71a018b73f127f8fc5c58ef3220b1d5335e7"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Dec 27 07:47:12 crc kubenswrapper[4934]: I1227 07:47:12.393051 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://5fa4c8b6a0bef1daea360a8dfb6c71a018b73f127f8fc5c58ef3220b1d5335e7" gracePeriod=30 Dec 27 07:47:12 crc kubenswrapper[4934]: I1227 07:47:12.558413 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 27 07:47:12 crc kubenswrapper[4934]: I1227 07:47:12.565267 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 27 07:47:12 crc kubenswrapper[4934]: I1227 07:47:12.682443 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 27 07:47:12 crc kubenswrapper[4934]: I1227 07:47:12.769512 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 27 07:47:12 crc kubenswrapper[4934]: I1227 07:47:12.770127 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 27 07:47:12 crc kubenswrapper[4934]: I1227 07:47:12.966176 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 27 07:47:12 crc kubenswrapper[4934]: I1227 07:47:12.976812 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 27 07:47:13 crc kubenswrapper[4934]: I1227 07:47:13.082978 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 27 07:47:13 crc kubenswrapper[4934]: I1227 07:47:13.187738 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 27 07:47:13 crc kubenswrapper[4934]: I1227 07:47:13.217001 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 27 07:47:13 crc kubenswrapper[4934]: I1227 07:47:13.257906 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 27 07:47:13 crc kubenswrapper[4934]: I1227 07:47:13.296840 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 27 07:47:13 crc kubenswrapper[4934]: I1227 07:47:13.315196 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 27 07:47:13 crc kubenswrapper[4934]: I1227 07:47:13.323832 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 27 07:47:13 crc kubenswrapper[4934]: I1227 07:47:13.422684 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 27 07:47:13 crc kubenswrapper[4934]: I1227 07:47:13.508524 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 27 07:47:13 crc kubenswrapper[4934]: I1227 07:47:13.688990 4934 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 27 07:47:13 crc kubenswrapper[4934]: I1227 07:47:13.860302 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 27 07:47:14 crc kubenswrapper[4934]: I1227 07:47:14.005914 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 27 07:47:14 crc kubenswrapper[4934]: I1227 07:47:14.165821 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 27 07:47:14 crc kubenswrapper[4934]: I1227 07:47:14.818620 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 27 07:47:14 crc kubenswrapper[4934]: I1227 07:47:14.957032 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 27 07:47:15 crc kubenswrapper[4934]: I1227 07:47:15.006212 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 27 07:47:15 crc kubenswrapper[4934]: I1227 07:47:15.060407 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 27 07:47:15 crc kubenswrapper[4934]: I1227 07:47:15.083732 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 27 07:47:18 crc kubenswrapper[4934]: I1227 07:47:18.261854 4934 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 27 07:47:18 crc kubenswrapper[4934]: I1227 07:47:18.262557 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://a4cbc79fe5abd8973450dc4962819f4db333ffbc34b1d459a9526114bc510ec5" gracePeriod=5 Dec 27 07:47:19 crc kubenswrapper[4934]: I1227 07:47:19.333016 4934 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.435926 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.436382 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.539007 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.539069 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.539106 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.539132 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.539196 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.539425 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.539419 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.539450 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.539479 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.545283 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.641695 4934 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.642182 4934 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.642259 4934 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.642322 4934 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.642385 4934 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.724547 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.724633 4934 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="a4cbc79fe5abd8973450dc4962819f4db333ffbc34b1d459a9526114bc510ec5" exitCode=137 Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.724692 4934 scope.go:117] "RemoveContainer" containerID="a4cbc79fe5abd8973450dc4962819f4db333ffbc34b1d459a9526114bc510ec5" Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.724725 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.749424 4934 scope.go:117] "RemoveContainer" containerID="a4cbc79fe5abd8973450dc4962819f4db333ffbc34b1d459a9526114bc510ec5" Dec 27 07:47:24 crc kubenswrapper[4934]: E1227 07:47:24.749960 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4cbc79fe5abd8973450dc4962819f4db333ffbc34b1d459a9526114bc510ec5\": container with ID starting with a4cbc79fe5abd8973450dc4962819f4db333ffbc34b1d459a9526114bc510ec5 not found: ID does not exist" containerID="a4cbc79fe5abd8973450dc4962819f4db333ffbc34b1d459a9526114bc510ec5" Dec 27 07:47:24 crc kubenswrapper[4934]: I1227 07:47:24.750069 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4cbc79fe5abd8973450dc4962819f4db333ffbc34b1d459a9526114bc510ec5"} err="failed to get container status \"a4cbc79fe5abd8973450dc4962819f4db333ffbc34b1d459a9526114bc510ec5\": rpc error: code = NotFound desc = could not find container \"a4cbc79fe5abd8973450dc4962819f4db333ffbc34b1d459a9526114bc510ec5\": container with ID starting with a4cbc79fe5abd8973450dc4962819f4db333ffbc34b1d459a9526114bc510ec5 not found: ID does not exist" Dec 27 07:47:25 crc kubenswrapper[4934]: I1227 07:47:25.474962 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 27 07:47:25 crc kubenswrapper[4934]: I1227 07:47:25.475349 4934 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 27 07:47:25 crc kubenswrapper[4934]: I1227 07:47:25.485323 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 27 07:47:25 crc kubenswrapper[4934]: I1227 07:47:25.485364 4934 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="da04b235-7dbc-4136-89dc-3c580d7155b3" Dec 27 07:47:25 crc kubenswrapper[4934]: I1227 07:47:25.490948 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 27 07:47:25 crc kubenswrapper[4934]: I1227 07:47:25.490983 4934 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="da04b235-7dbc-4136-89dc-3c580d7155b3" Dec 27 07:47:42 crc kubenswrapper[4934]: I1227 07:47:42.842956 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 27 07:47:42 crc kubenswrapper[4934]: I1227 07:47:42.846180 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 27 07:47:42 crc kubenswrapper[4934]: I1227 07:47:42.846259 4934 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="5fa4c8b6a0bef1daea360a8dfb6c71a018b73f127f8fc5c58ef3220b1d5335e7" exitCode=137 Dec 27 07:47:42 crc kubenswrapper[4934]: I1227 07:47:42.846306 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"5fa4c8b6a0bef1daea360a8dfb6c71a018b73f127f8fc5c58ef3220b1d5335e7"} Dec 27 07:47:42 crc kubenswrapper[4934]: I1227 07:47:42.846366 4934 scope.go:117] "RemoveContainer" containerID="bd5f3efc15d92f33024204b2490f6964847bbf7364dc5da8b28141c6a84fcd47" Dec 27 07:47:43 crc kubenswrapper[4934]: I1227 07:47:43.855216 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 27 07:47:43 crc kubenswrapper[4934]: I1227 07:47:43.856717 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6d66e1134058289fe715318d0983996f62e222c539a7a940eb1ab6feb1889204"} Dec 27 07:47:46 crc kubenswrapper[4934]: I1227 07:47:46.353699 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:47:52 crc kubenswrapper[4934]: I1227 07:47:52.392191 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:47:52 crc kubenswrapper[4934]: I1227 07:47:52.402723 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:47:56 crc kubenswrapper[4934]: I1227 07:47:56.360795 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.078217 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-d28xm"] Dec 27 07:48:02 crc kubenswrapper[4934]: E1227 07:48:02.079001 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52a38d21-a579-4f37-be2f-856e2f837f83" containerName="extract-content" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079016 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="52a38d21-a579-4f37-be2f-856e2f837f83" containerName="extract-content" Dec 27 07:48:02 crc kubenswrapper[4934]: E1227 07:48:02.079032 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" containerName="extract-utilities" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079042 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" containerName="extract-utilities" Dec 27 07:48:02 crc kubenswrapper[4934]: E1227 07:48:02.079053 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b06ee67d-88e4-4c2f-b0a1-efdc7517c40d" containerName="registry-server" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079061 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b06ee67d-88e4-4c2f-b0a1-efdc7517c40d" containerName="registry-server" Dec 27 07:48:02 crc kubenswrapper[4934]: E1227 07:48:02.079073 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079105 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 27 07:48:02 crc kubenswrapper[4934]: E1227 07:48:02.079117 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deaa2d15-bcaf-4568-8847-7bccae9205e0" containerName="extract-content" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079123 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="deaa2d15-bcaf-4568-8847-7bccae9205e0" containerName="extract-content" Dec 27 07:48:02 crc kubenswrapper[4934]: E1227 07:48:02.079135 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8ce8d56-fde4-44c6-a3df-53d4dbca22d4" containerName="installer" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079142 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8ce8d56-fde4-44c6-a3df-53d4dbca22d4" containerName="installer" Dec 27 07:48:02 crc kubenswrapper[4934]: E1227 07:48:02.079153 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deaa2d15-bcaf-4568-8847-7bccae9205e0" containerName="registry-server" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079169 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="deaa2d15-bcaf-4568-8847-7bccae9205e0" containerName="registry-server" Dec 27 07:48:02 crc kubenswrapper[4934]: E1227 07:48:02.079181 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deaa2d15-bcaf-4568-8847-7bccae9205e0" containerName="extract-utilities" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079188 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="deaa2d15-bcaf-4568-8847-7bccae9205e0" containerName="extract-utilities" Dec 27 07:48:02 crc kubenswrapper[4934]: E1227 07:48:02.079199 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b06ee67d-88e4-4c2f-b0a1-efdc7517c40d" containerName="extract-content" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079207 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b06ee67d-88e4-4c2f-b0a1-efdc7517c40d" containerName="extract-content" Dec 27 07:48:02 crc kubenswrapper[4934]: E1227 07:48:02.079217 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52a38d21-a579-4f37-be2f-856e2f837f83" containerName="extract-utilities" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079224 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="52a38d21-a579-4f37-be2f-856e2f837f83" containerName="extract-utilities" Dec 27 07:48:02 crc kubenswrapper[4934]: E1227 07:48:02.079234 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19b07e13-aa5d-4a0a-b6d9-254afc234025" containerName="extract-content" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079242 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="19b07e13-aa5d-4a0a-b6d9-254afc234025" containerName="extract-content" Dec 27 07:48:02 crc kubenswrapper[4934]: E1227 07:48:02.079252 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b06ee67d-88e4-4c2f-b0a1-efdc7517c40d" containerName="extract-utilities" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079261 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b06ee67d-88e4-4c2f-b0a1-efdc7517c40d" containerName="extract-utilities" Dec 27 07:48:02 crc kubenswrapper[4934]: E1227 07:48:02.079268 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" containerName="extract-content" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079294 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" containerName="extract-content" Dec 27 07:48:02 crc kubenswrapper[4934]: E1227 07:48:02.079303 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" containerName="registry-server" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079310 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" containerName="registry-server" Dec 27 07:48:02 crc kubenswrapper[4934]: E1227 07:48:02.079326 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="941a07b0-8984-43f4-8fcd-6cdc8c88f14a" containerName="marketplace-operator" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079334 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="941a07b0-8984-43f4-8fcd-6cdc8c88f14a" containerName="marketplace-operator" Dec 27 07:48:02 crc kubenswrapper[4934]: E1227 07:48:02.079342 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19b07e13-aa5d-4a0a-b6d9-254afc234025" containerName="registry-server" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079349 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="19b07e13-aa5d-4a0a-b6d9-254afc234025" containerName="registry-server" Dec 27 07:48:02 crc kubenswrapper[4934]: E1227 07:48:02.079358 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52a38d21-a579-4f37-be2f-856e2f837f83" containerName="registry-server" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079368 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="52a38d21-a579-4f37-be2f-856e2f837f83" containerName="registry-server" Dec 27 07:48:02 crc kubenswrapper[4934]: E1227 07:48:02.079378 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19b07e13-aa5d-4a0a-b6d9-254afc234025" containerName="extract-utilities" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079386 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="19b07e13-aa5d-4a0a-b6d9-254afc234025" containerName="extract-utilities" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079495 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="deaa2d15-bcaf-4568-8847-7bccae9205e0" containerName="registry-server" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079505 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079517 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccaf0173-eafb-4a76-bdd8-4a9bb1dfdd20" containerName="registry-server" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079529 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="52a38d21-a579-4f37-be2f-856e2f837f83" containerName="registry-server" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079540 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8ce8d56-fde4-44c6-a3df-53d4dbca22d4" containerName="installer" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079555 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="b06ee67d-88e4-4c2f-b0a1-efdc7517c40d" containerName="registry-server" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079564 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="941a07b0-8984-43f4-8fcd-6cdc8c88f14a" containerName="marketplace-operator" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.079572 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="19b07e13-aa5d-4a0a-b6d9-254afc234025" containerName="registry-server" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.080022 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-d28xm" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.082268 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5jzhr"] Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.082892 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.084635 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"openshift-service-ca.crt" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.085104 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemetry-config" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.085289 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-dockercfg-wwt9l" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.085337 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.092423 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-tls" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.092493 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.092585 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-root-ca.crt" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.093305 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.093342 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.109123 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.110484 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-j5cpz"] Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.115206 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5jzhr"] Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.116982 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-d28xm"] Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.121035 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbshx\" (UniqueName: \"kubernetes.io/projected/8bc955ec-6c75-4f54-b07b-78334f79477f-kube-api-access-qbshx\") pod \"cluster-monitoring-operator-6d5b84845-d28xm\" (UID: \"8bc955ec-6c75-4f54-b07b-78334f79477f\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-d28xm" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.121095 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a4612dd4-0e0d-4c38-8da5-8bc30dee7c12-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5jzhr\" (UID: \"a4612dd4-0e0d-4c38-8da5-8bc30dee7c12\") " pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.121121 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a4612dd4-0e0d-4c38-8da5-8bc30dee7c12-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5jzhr\" (UID: \"a4612dd4-0e0d-4c38-8da5-8bc30dee7c12\") " pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.121182 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vc5nr\" (UniqueName: \"kubernetes.io/projected/a4612dd4-0e0d-4c38-8da5-8bc30dee7c12-kube-api-access-vc5nr\") pod \"marketplace-operator-79b997595-5jzhr\" (UID: \"a4612dd4-0e0d-4c38-8da5-8bc30dee7c12\") " pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.121205 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/8bc955ec-6c75-4f54-b07b-78334f79477f-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-d28xm\" (UID: \"8bc955ec-6c75-4f54-b07b-78334f79477f\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-d28xm" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.121231 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/8bc955ec-6c75-4f54-b07b-78334f79477f-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-d28xm\" (UID: \"8bc955ec-6c75-4f54-b07b-78334f79477f\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-d28xm" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.155975 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w"] Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.156606 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" podUID="32619f39-2bc4-4208-bd1d-9d715cf2ad09" containerName="controller-manager" containerID="cri-o://1b261a8503d149ca4b20616f4a66b4714f8e7c85bd4380fd3a6570572fd8cf41" gracePeriod=30 Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.187271 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj"] Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.187500 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" podUID="f89e5ac0-bc03-4af9-91bb-918412fa9b47" containerName="route-controller-manager" containerID="cri-o://fa149a7dc9e65f9245738329973d99cf95574f29899cdb8b9bfbe54a63964440" gracePeriod=30 Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.222865 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a4612dd4-0e0d-4c38-8da5-8bc30dee7c12-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5jzhr\" (UID: \"a4612dd4-0e0d-4c38-8da5-8bc30dee7c12\") " pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.222924 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a4612dd4-0e0d-4c38-8da5-8bc30dee7c12-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5jzhr\" (UID: \"a4612dd4-0e0d-4c38-8da5-8bc30dee7c12\") " pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.222991 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vc5nr\" (UniqueName: \"kubernetes.io/projected/a4612dd4-0e0d-4c38-8da5-8bc30dee7c12-kube-api-access-vc5nr\") pod \"marketplace-operator-79b997595-5jzhr\" (UID: \"a4612dd4-0e0d-4c38-8da5-8bc30dee7c12\") " pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.223019 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/8bc955ec-6c75-4f54-b07b-78334f79477f-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-d28xm\" (UID: \"8bc955ec-6c75-4f54-b07b-78334f79477f\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-d28xm" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.223050 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/8bc955ec-6c75-4f54-b07b-78334f79477f-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-d28xm\" (UID: \"8bc955ec-6c75-4f54-b07b-78334f79477f\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-d28xm" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.223112 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbshx\" (UniqueName: \"kubernetes.io/projected/8bc955ec-6c75-4f54-b07b-78334f79477f-kube-api-access-qbshx\") pod \"cluster-monitoring-operator-6d5b84845-d28xm\" (UID: \"8bc955ec-6c75-4f54-b07b-78334f79477f\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-d28xm" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.224338 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/8bc955ec-6c75-4f54-b07b-78334f79477f-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-d28xm\" (UID: \"8bc955ec-6c75-4f54-b07b-78334f79477f\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-d28xm" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.224828 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a4612dd4-0e0d-4c38-8da5-8bc30dee7c12-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5jzhr\" (UID: \"a4612dd4-0e0d-4c38-8da5-8bc30dee7c12\") " pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.233740 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a4612dd4-0e0d-4c38-8da5-8bc30dee7c12-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5jzhr\" (UID: \"a4612dd4-0e0d-4c38-8da5-8bc30dee7c12\") " pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.248007 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbshx\" (UniqueName: \"kubernetes.io/projected/8bc955ec-6c75-4f54-b07b-78334f79477f-kube-api-access-qbshx\") pod \"cluster-monitoring-operator-6d5b84845-d28xm\" (UID: \"8bc955ec-6c75-4f54-b07b-78334f79477f\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-d28xm" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.248698 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/8bc955ec-6c75-4f54-b07b-78334f79477f-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-d28xm\" (UID: \"8bc955ec-6c75-4f54-b07b-78334f79477f\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-d28xm" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.253829 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vc5nr\" (UniqueName: \"kubernetes.io/projected/a4612dd4-0e0d-4c38-8da5-8bc30dee7c12-kube-api-access-vc5nr\") pod \"marketplace-operator-79b997595-5jzhr\" (UID: \"a4612dd4-0e0d-4c38-8da5-8bc30dee7c12\") " pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.443605 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-d28xm" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.463940 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.692934 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.734160 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.778539 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5jzhr"] Dec 27 07:48:02 crc kubenswrapper[4934]: W1227 07:48:02.789160 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4612dd4_0e0d_4c38_8da5_8bc30dee7c12.slice/crio-214275c8015169be5e45846518c78d07f35c1b0b4fe440f82cb4deaf9a4187b0 WatchSource:0}: Error finding container 214275c8015169be5e45846518c78d07f35c1b0b4fe440f82cb4deaf9a4187b0: Status 404 returned error can't find the container with id 214275c8015169be5e45846518c78d07f35c1b0b4fe440f82cb4deaf9a4187b0 Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.831693 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f89e5ac0-bc03-4af9-91bb-918412fa9b47-config\") pod \"f89e5ac0-bc03-4af9-91bb-918412fa9b47\" (UID: \"f89e5ac0-bc03-4af9-91bb-918412fa9b47\") " Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.832038 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32619f39-2bc4-4208-bd1d-9d715cf2ad09-serving-cert\") pod \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\" (UID: \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\") " Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.832073 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f89e5ac0-bc03-4af9-91bb-918412fa9b47-client-ca\") pod \"f89e5ac0-bc03-4af9-91bb-918412fa9b47\" (UID: \"f89e5ac0-bc03-4af9-91bb-918412fa9b47\") " Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.832118 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/32619f39-2bc4-4208-bd1d-9d715cf2ad09-proxy-ca-bundles\") pod \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\" (UID: \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\") " Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.832163 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32619f39-2bc4-4208-bd1d-9d715cf2ad09-client-ca\") pod \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\" (UID: \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\") " Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.832195 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f89e5ac0-bc03-4af9-91bb-918412fa9b47-serving-cert\") pod \"f89e5ac0-bc03-4af9-91bb-918412fa9b47\" (UID: \"f89e5ac0-bc03-4af9-91bb-918412fa9b47\") " Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.832235 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtndr\" (UniqueName: \"kubernetes.io/projected/32619f39-2bc4-4208-bd1d-9d715cf2ad09-kube-api-access-gtndr\") pod \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\" (UID: \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\") " Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.832281 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32619f39-2bc4-4208-bd1d-9d715cf2ad09-config\") pod \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\" (UID: \"32619f39-2bc4-4208-bd1d-9d715cf2ad09\") " Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.832308 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72bd5\" (UniqueName: \"kubernetes.io/projected/f89e5ac0-bc03-4af9-91bb-918412fa9b47-kube-api-access-72bd5\") pod \"f89e5ac0-bc03-4af9-91bb-918412fa9b47\" (UID: \"f89e5ac0-bc03-4af9-91bb-918412fa9b47\") " Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.834298 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f89e5ac0-bc03-4af9-91bb-918412fa9b47-client-ca" (OuterVolumeSpecName: "client-ca") pod "f89e5ac0-bc03-4af9-91bb-918412fa9b47" (UID: "f89e5ac0-bc03-4af9-91bb-918412fa9b47"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.834879 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32619f39-2bc4-4208-bd1d-9d715cf2ad09-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "32619f39-2bc4-4208-bd1d-9d715cf2ad09" (UID: "32619f39-2bc4-4208-bd1d-9d715cf2ad09"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.836026 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32619f39-2bc4-4208-bd1d-9d715cf2ad09-config" (OuterVolumeSpecName: "config") pod "32619f39-2bc4-4208-bd1d-9d715cf2ad09" (UID: "32619f39-2bc4-4208-bd1d-9d715cf2ad09"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.836282 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32619f39-2bc4-4208-bd1d-9d715cf2ad09-client-ca" (OuterVolumeSpecName: "client-ca") pod "32619f39-2bc4-4208-bd1d-9d715cf2ad09" (UID: "32619f39-2bc4-4208-bd1d-9d715cf2ad09"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.836555 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f89e5ac0-bc03-4af9-91bb-918412fa9b47-config" (OuterVolumeSpecName: "config") pod "f89e5ac0-bc03-4af9-91bb-918412fa9b47" (UID: "f89e5ac0-bc03-4af9-91bb-918412fa9b47"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.837591 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32619f39-2bc4-4208-bd1d-9d715cf2ad09-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "32619f39-2bc4-4208-bd1d-9d715cf2ad09" (UID: "32619f39-2bc4-4208-bd1d-9d715cf2ad09"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.837737 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f89e5ac0-bc03-4af9-91bb-918412fa9b47-kube-api-access-72bd5" (OuterVolumeSpecName: "kube-api-access-72bd5") pod "f89e5ac0-bc03-4af9-91bb-918412fa9b47" (UID: "f89e5ac0-bc03-4af9-91bb-918412fa9b47"). InnerVolumeSpecName "kube-api-access-72bd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.837788 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f89e5ac0-bc03-4af9-91bb-918412fa9b47-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f89e5ac0-bc03-4af9-91bb-918412fa9b47" (UID: "f89e5ac0-bc03-4af9-91bb-918412fa9b47"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.837911 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32619f39-2bc4-4208-bd1d-9d715cf2ad09-kube-api-access-gtndr" (OuterVolumeSpecName: "kube-api-access-gtndr") pod "32619f39-2bc4-4208-bd1d-9d715cf2ad09" (UID: "32619f39-2bc4-4208-bd1d-9d715cf2ad09"). InnerVolumeSpecName "kube-api-access-gtndr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.892140 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-d28xm"] Dec 27 07:48:02 crc kubenswrapper[4934]: W1227 07:48:02.894180 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8bc955ec_6c75_4f54_b07b_78334f79477f.slice/crio-cade86034f16b58153f98e99973aff977070165eadc73c443ad7d2ff2fbb9974 WatchSource:0}: Error finding container cade86034f16b58153f98e99973aff977070165eadc73c443ad7d2ff2fbb9974: Status 404 returned error can't find the container with id cade86034f16b58153f98e99973aff977070165eadc73c443ad7d2ff2fbb9974 Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.934259 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f89e5ac0-bc03-4af9-91bb-918412fa9b47-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.934299 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/32619f39-2bc4-4208-bd1d-9d715cf2ad09-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.934313 4934 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f89e5ac0-bc03-4af9-91bb-918412fa9b47-client-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.934326 4934 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/32619f39-2bc4-4208-bd1d-9d715cf2ad09-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.934341 4934 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/32619f39-2bc4-4208-bd1d-9d715cf2ad09-client-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.934352 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f89e5ac0-bc03-4af9-91bb-918412fa9b47-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.934366 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtndr\" (UniqueName: \"kubernetes.io/projected/32619f39-2bc4-4208-bd1d-9d715cf2ad09-kube-api-access-gtndr\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.934378 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/32619f39-2bc4-4208-bd1d-9d715cf2ad09-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.934390 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72bd5\" (UniqueName: \"kubernetes.io/projected/f89e5ac0-bc03-4af9-91bb-918412fa9b47-kube-api-access-72bd5\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.957711 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-d28xm" event={"ID":"8bc955ec-6c75-4f54-b07b-78334f79477f","Type":"ContainerStarted","Data":"cade86034f16b58153f98e99973aff977070165eadc73c443ad7d2ff2fbb9974"} Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.959471 4934 generic.go:334] "Generic (PLEG): container finished" podID="32619f39-2bc4-4208-bd1d-9d715cf2ad09" containerID="1b261a8503d149ca4b20616f4a66b4714f8e7c85bd4380fd3a6570572fd8cf41" exitCode=0 Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.959515 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" event={"ID":"32619f39-2bc4-4208-bd1d-9d715cf2ad09","Type":"ContainerDied","Data":"1b261a8503d149ca4b20616f4a66b4714f8e7c85bd4380fd3a6570572fd8cf41"} Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.959560 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" event={"ID":"32619f39-2bc4-4208-bd1d-9d715cf2ad09","Type":"ContainerDied","Data":"b67a8ba7d01ca1cc88079b87f884cbac1b8fe5de1f1892cde8216eab64d8009c"} Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.959526 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.959583 4934 scope.go:117] "RemoveContainer" containerID="1b261a8503d149ca4b20616f4a66b4714f8e7c85bd4380fd3a6570572fd8cf41" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.961867 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" event={"ID":"a4612dd4-0e0d-4c38-8da5-8bc30dee7c12","Type":"ContainerStarted","Data":"214275c8015169be5e45846518c78d07f35c1b0b4fe440f82cb4deaf9a4187b0"} Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.965474 4934 generic.go:334] "Generic (PLEG): container finished" podID="f89e5ac0-bc03-4af9-91bb-918412fa9b47" containerID="fa149a7dc9e65f9245738329973d99cf95574f29899cdb8b9bfbe54a63964440" exitCode=0 Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.965509 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" event={"ID":"f89e5ac0-bc03-4af9-91bb-918412fa9b47","Type":"ContainerDied","Data":"fa149a7dc9e65f9245738329973d99cf95574f29899cdb8b9bfbe54a63964440"} Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.965528 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" event={"ID":"f89e5ac0-bc03-4af9-91bb-918412fa9b47","Type":"ContainerDied","Data":"fe76c6bc684aa45ccbbcf5c35202afc840d77c96fc096bd6c159532242204bff"} Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.965582 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.983679 4934 scope.go:117] "RemoveContainer" containerID="1b261a8503d149ca4b20616f4a66b4714f8e7c85bd4380fd3a6570572fd8cf41" Dec 27 07:48:02 crc kubenswrapper[4934]: E1227 07:48:02.984130 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b261a8503d149ca4b20616f4a66b4714f8e7c85bd4380fd3a6570572fd8cf41\": container with ID starting with 1b261a8503d149ca4b20616f4a66b4714f8e7c85bd4380fd3a6570572fd8cf41 not found: ID does not exist" containerID="1b261a8503d149ca4b20616f4a66b4714f8e7c85bd4380fd3a6570572fd8cf41" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.984173 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b261a8503d149ca4b20616f4a66b4714f8e7c85bd4380fd3a6570572fd8cf41"} err="failed to get container status \"1b261a8503d149ca4b20616f4a66b4714f8e7c85bd4380fd3a6570572fd8cf41\": rpc error: code = NotFound desc = could not find container \"1b261a8503d149ca4b20616f4a66b4714f8e7c85bd4380fd3a6570572fd8cf41\": container with ID starting with 1b261a8503d149ca4b20616f4a66b4714f8e7c85bd4380fd3a6570572fd8cf41 not found: ID does not exist" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.984202 4934 scope.go:117] "RemoveContainer" containerID="fa149a7dc9e65f9245738329973d99cf95574f29899cdb8b9bfbe54a63964440" Dec 27 07:48:02 crc kubenswrapper[4934]: I1227 07:48:02.994626 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w"] Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.002279 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5d6b58cf7d-hgq5w"] Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.006245 4934 scope.go:117] "RemoveContainer" containerID="fa149a7dc9e65f9245738329973d99cf95574f29899cdb8b9bfbe54a63964440" Dec 27 07:48:03 crc kubenswrapper[4934]: E1227 07:48:03.006693 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa149a7dc9e65f9245738329973d99cf95574f29899cdb8b9bfbe54a63964440\": container with ID starting with fa149a7dc9e65f9245738329973d99cf95574f29899cdb8b9bfbe54a63964440 not found: ID does not exist" containerID="fa149a7dc9e65f9245738329973d99cf95574f29899cdb8b9bfbe54a63964440" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.006732 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa149a7dc9e65f9245738329973d99cf95574f29899cdb8b9bfbe54a63964440"} err="failed to get container status \"fa149a7dc9e65f9245738329973d99cf95574f29899cdb8b9bfbe54a63964440\": rpc error: code = NotFound desc = could not find container \"fa149a7dc9e65f9245738329973d99cf95574f29899cdb8b9bfbe54a63964440\": container with ID starting with fa149a7dc9e65f9245738329973d99cf95574f29899cdb8b9bfbe54a63964440 not found: ID does not exist" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.007832 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj"] Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.012357 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58564d585f-nmxfj"] Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.478023 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32619f39-2bc4-4208-bd1d-9d715cf2ad09" path="/var/lib/kubelet/pods/32619f39-2bc4-4208-bd1d-9d715cf2ad09/volumes" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.479921 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f89e5ac0-bc03-4af9-91bb-918412fa9b47" path="/var/lib/kubelet/pods/f89e5ac0-bc03-4af9-91bb-918412fa9b47/volumes" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.676124 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-589f9f9656-pbctv"] Dec 27 07:48:03 crc kubenswrapper[4934]: E1227 07:48:03.676500 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32619f39-2bc4-4208-bd1d-9d715cf2ad09" containerName="controller-manager" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.676518 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="32619f39-2bc4-4208-bd1d-9d715cf2ad09" containerName="controller-manager" Dec 27 07:48:03 crc kubenswrapper[4934]: E1227 07:48:03.676583 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f89e5ac0-bc03-4af9-91bb-918412fa9b47" containerName="route-controller-manager" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.676869 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f89e5ac0-bc03-4af9-91bb-918412fa9b47" containerName="route-controller-manager" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.676976 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f89e5ac0-bc03-4af9-91bb-918412fa9b47" containerName="route-controller-manager" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.676989 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="32619f39-2bc4-4208-bd1d-9d715cf2ad09" containerName="controller-manager" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.677471 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.680719 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb"] Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.681483 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.682517 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.682807 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.685414 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.687409 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.687625 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.687802 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.688380 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.688684 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.688759 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.689029 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.689314 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.689524 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.696713 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.700152 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb"] Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.706987 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-589f9f9656-pbctv"] Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.743822 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4598b565-6f32-486e-83ad-f72b288b5a38-serving-cert\") pod \"route-controller-manager-5fdfb79994-qlktb\" (UID: \"4598b565-6f32-486e-83ad-f72b288b5a38\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.743877 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4598b565-6f32-486e-83ad-f72b288b5a38-config\") pod \"route-controller-manager-5fdfb79994-qlktb\" (UID: \"4598b565-6f32-486e-83ad-f72b288b5a38\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.743905 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47vjv\" (UniqueName: \"kubernetes.io/projected/46338877-dff9-4023-b399-212fce6fae21-kube-api-access-47vjv\") pod \"controller-manager-589f9f9656-pbctv\" (UID: \"46338877-dff9-4023-b399-212fce6fae21\") " pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.743933 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/46338877-dff9-4023-b399-212fce6fae21-client-ca\") pod \"controller-manager-589f9f9656-pbctv\" (UID: \"46338877-dff9-4023-b399-212fce6fae21\") " pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.744067 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9mtm\" (UniqueName: \"kubernetes.io/projected/4598b565-6f32-486e-83ad-f72b288b5a38-kube-api-access-c9mtm\") pod \"route-controller-manager-5fdfb79994-qlktb\" (UID: \"4598b565-6f32-486e-83ad-f72b288b5a38\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.744183 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46338877-dff9-4023-b399-212fce6fae21-config\") pod \"controller-manager-589f9f9656-pbctv\" (UID: \"46338877-dff9-4023-b399-212fce6fae21\") " pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.744254 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4598b565-6f32-486e-83ad-f72b288b5a38-client-ca\") pod \"route-controller-manager-5fdfb79994-qlktb\" (UID: \"4598b565-6f32-486e-83ad-f72b288b5a38\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.744289 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46338877-dff9-4023-b399-212fce6fae21-serving-cert\") pod \"controller-manager-589f9f9656-pbctv\" (UID: \"46338877-dff9-4023-b399-212fce6fae21\") " pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.744313 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/46338877-dff9-4023-b399-212fce6fae21-proxy-ca-bundles\") pod \"controller-manager-589f9f9656-pbctv\" (UID: \"46338877-dff9-4023-b399-212fce6fae21\") " pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.845483 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4598b565-6f32-486e-83ad-f72b288b5a38-config\") pod \"route-controller-manager-5fdfb79994-qlktb\" (UID: \"4598b565-6f32-486e-83ad-f72b288b5a38\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.845544 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47vjv\" (UniqueName: \"kubernetes.io/projected/46338877-dff9-4023-b399-212fce6fae21-kube-api-access-47vjv\") pod \"controller-manager-589f9f9656-pbctv\" (UID: \"46338877-dff9-4023-b399-212fce6fae21\") " pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.845584 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/46338877-dff9-4023-b399-212fce6fae21-client-ca\") pod \"controller-manager-589f9f9656-pbctv\" (UID: \"46338877-dff9-4023-b399-212fce6fae21\") " pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.845629 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9mtm\" (UniqueName: \"kubernetes.io/projected/4598b565-6f32-486e-83ad-f72b288b5a38-kube-api-access-c9mtm\") pod \"route-controller-manager-5fdfb79994-qlktb\" (UID: \"4598b565-6f32-486e-83ad-f72b288b5a38\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.845678 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46338877-dff9-4023-b399-212fce6fae21-config\") pod \"controller-manager-589f9f9656-pbctv\" (UID: \"46338877-dff9-4023-b399-212fce6fae21\") " pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.845713 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4598b565-6f32-486e-83ad-f72b288b5a38-client-ca\") pod \"route-controller-manager-5fdfb79994-qlktb\" (UID: \"4598b565-6f32-486e-83ad-f72b288b5a38\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.845733 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46338877-dff9-4023-b399-212fce6fae21-serving-cert\") pod \"controller-manager-589f9f9656-pbctv\" (UID: \"46338877-dff9-4023-b399-212fce6fae21\") " pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.845757 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/46338877-dff9-4023-b399-212fce6fae21-proxy-ca-bundles\") pod \"controller-manager-589f9f9656-pbctv\" (UID: \"46338877-dff9-4023-b399-212fce6fae21\") " pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.845783 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4598b565-6f32-486e-83ad-f72b288b5a38-serving-cert\") pod \"route-controller-manager-5fdfb79994-qlktb\" (UID: \"4598b565-6f32-486e-83ad-f72b288b5a38\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.846600 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/46338877-dff9-4023-b399-212fce6fae21-client-ca\") pod \"controller-manager-589f9f9656-pbctv\" (UID: \"46338877-dff9-4023-b399-212fce6fae21\") " pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.846894 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4598b565-6f32-486e-83ad-f72b288b5a38-client-ca\") pod \"route-controller-manager-5fdfb79994-qlktb\" (UID: \"4598b565-6f32-486e-83ad-f72b288b5a38\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.847352 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4598b565-6f32-486e-83ad-f72b288b5a38-config\") pod \"route-controller-manager-5fdfb79994-qlktb\" (UID: \"4598b565-6f32-486e-83ad-f72b288b5a38\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.847458 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46338877-dff9-4023-b399-212fce6fae21-config\") pod \"controller-manager-589f9f9656-pbctv\" (UID: \"46338877-dff9-4023-b399-212fce6fae21\") " pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.847526 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/46338877-dff9-4023-b399-212fce6fae21-proxy-ca-bundles\") pod \"controller-manager-589f9f9656-pbctv\" (UID: \"46338877-dff9-4023-b399-212fce6fae21\") " pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.850791 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4598b565-6f32-486e-83ad-f72b288b5a38-serving-cert\") pod \"route-controller-manager-5fdfb79994-qlktb\" (UID: \"4598b565-6f32-486e-83ad-f72b288b5a38\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.866878 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46338877-dff9-4023-b399-212fce6fae21-serving-cert\") pod \"controller-manager-589f9f9656-pbctv\" (UID: \"46338877-dff9-4023-b399-212fce6fae21\") " pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.869771 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47vjv\" (UniqueName: \"kubernetes.io/projected/46338877-dff9-4023-b399-212fce6fae21-kube-api-access-47vjv\") pod \"controller-manager-589f9f9656-pbctv\" (UID: \"46338877-dff9-4023-b399-212fce6fae21\") " pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.869768 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9mtm\" (UniqueName: \"kubernetes.io/projected/4598b565-6f32-486e-83ad-f72b288b5a38-kube-api-access-c9mtm\") pod \"route-controller-manager-5fdfb79994-qlktb\" (UID: \"4598b565-6f32-486e-83ad-f72b288b5a38\") " pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.975434 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" event={"ID":"a4612dd4-0e0d-4c38-8da5-8bc30dee7c12","Type":"ContainerStarted","Data":"67e9d2fa2404b9f860a2bbc81d950b59caee85958a0d2750c45a0fcbd374a629"} Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.975659 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.983032 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" Dec 27 07:48:03 crc kubenswrapper[4934]: I1227 07:48:03.993818 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" podStartSLOduration=1.9937984260000001 podStartE2EDuration="1.993798426s" podCreationTimestamp="2025-12-27 07:48:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:48:03.990430622 +0000 UTC m=+344.810871306" watchObservedRunningTime="2025-12-27 07:48:03.993798426 +0000 UTC m=+344.814239020" Dec 27 07:48:04 crc kubenswrapper[4934]: I1227 07:48:04.004303 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" Dec 27 07:48:04 crc kubenswrapper[4934]: I1227 07:48:04.014182 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" Dec 27 07:48:04 crc kubenswrapper[4934]: I1227 07:48:04.232250 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb"] Dec 27 07:48:04 crc kubenswrapper[4934]: W1227 07:48:04.242071 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4598b565_6f32_486e_83ad_f72b288b5a38.slice/crio-af428d12d26c7d9e74deae0f6995710ca1ee9bab4c22022b8abef66bb340aa4a WatchSource:0}: Error finding container af428d12d26c7d9e74deae0f6995710ca1ee9bab4c22022b8abef66bb340aa4a: Status 404 returned error can't find the container with id af428d12d26c7d9e74deae0f6995710ca1ee9bab4c22022b8abef66bb340aa4a Dec 27 07:48:04 crc kubenswrapper[4934]: I1227 07:48:04.566677 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-589f9f9656-pbctv"] Dec 27 07:48:04 crc kubenswrapper[4934]: I1227 07:48:04.981005 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" event={"ID":"4598b565-6f32-486e-83ad-f72b288b5a38","Type":"ContainerStarted","Data":"4750e5a042292fc070216f74a477b0be4a23812c65aca87e5ef8c6f02a567d35"} Dec 27 07:48:04 crc kubenswrapper[4934]: I1227 07:48:04.981047 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" event={"ID":"4598b565-6f32-486e-83ad-f72b288b5a38","Type":"ContainerStarted","Data":"af428d12d26c7d9e74deae0f6995710ca1ee9bab4c22022b8abef66bb340aa4a"} Dec 27 07:48:05 crc kubenswrapper[4934]: I1227 07:48:05.000969 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" podStartSLOduration=3.000953771 podStartE2EDuration="3.000953771s" podCreationTimestamp="2025-12-27 07:48:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:48:04.997072695 +0000 UTC m=+345.817513299" watchObservedRunningTime="2025-12-27 07:48:05.000953771 +0000 UTC m=+345.821394365" Dec 27 07:48:05 crc kubenswrapper[4934]: W1227 07:48:05.063602 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46338877_dff9_4023_b399_212fce6fae21.slice/crio-3059cdf7a39b73ce325003d215e8d003ac73ffa9603987f2fe77af21ace3b80e WatchSource:0}: Error finding container 3059cdf7a39b73ce325003d215e8d003ac73ffa9603987f2fe77af21ace3b80e: Status 404 returned error can't find the container with id 3059cdf7a39b73ce325003d215e8d003ac73ffa9603987f2fe77af21ace3b80e Dec 27 07:48:05 crc kubenswrapper[4934]: I1227 07:48:05.695773 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2"] Dec 27 07:48:05 crc kubenswrapper[4934]: I1227 07:48:05.696353 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" Dec 27 07:48:05 crc kubenswrapper[4934]: I1227 07:48:05.698196 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-tls" Dec 27 07:48:05 crc kubenswrapper[4934]: I1227 07:48:05.704674 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2"] Dec 27 07:48:05 crc kubenswrapper[4934]: I1227 07:48:05.765156 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/964b7818-b04e-4c6d-8ad0-5a108513be73-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-2n6w2\" (UID: \"964b7818-b04e-4c6d-8ad0-5a108513be73\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" Dec 27 07:48:05 crc kubenswrapper[4934]: I1227 07:48:05.865942 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/964b7818-b04e-4c6d-8ad0-5a108513be73-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-2n6w2\" (UID: \"964b7818-b04e-4c6d-8ad0-5a108513be73\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" Dec 27 07:48:05 crc kubenswrapper[4934]: E1227 07:48:05.866173 4934 secret.go:188] Couldn't get secret openshift-monitoring/prometheus-operator-admission-webhook-tls: secret "prometheus-operator-admission-webhook-tls" not found Dec 27 07:48:05 crc kubenswrapper[4934]: E1227 07:48:05.866261 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/964b7818-b04e-4c6d-8ad0-5a108513be73-tls-certificates podName:964b7818-b04e-4c6d-8ad0-5a108513be73 nodeName:}" failed. No retries permitted until 2025-12-27 07:48:06.366236004 +0000 UTC m=+347.186676588 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-certificates" (UniqueName: "kubernetes.io/secret/964b7818-b04e-4c6d-8ad0-5a108513be73-tls-certificates") pod "prometheus-operator-admission-webhook-f54c54754-2n6w2" (UID: "964b7818-b04e-4c6d-8ad0-5a108513be73") : secret "prometheus-operator-admission-webhook-tls" not found Dec 27 07:48:05 crc kubenswrapper[4934]: I1227 07:48:05.989778 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" event={"ID":"46338877-dff9-4023-b399-212fce6fae21","Type":"ContainerStarted","Data":"bad748b3be090ee97861e441494efe15616c155cc762df3c3fad147ecb21ed08"} Dec 27 07:48:05 crc kubenswrapper[4934]: I1227 07:48:05.989829 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" event={"ID":"46338877-dff9-4023-b399-212fce6fae21","Type":"ContainerStarted","Data":"3059cdf7a39b73ce325003d215e8d003ac73ffa9603987f2fe77af21ace3b80e"} Dec 27 07:48:05 crc kubenswrapper[4934]: I1227 07:48:05.990164 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" Dec 27 07:48:05 crc kubenswrapper[4934]: I1227 07:48:05.993523 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-d28xm" event={"ID":"8bc955ec-6c75-4f54-b07b-78334f79477f","Type":"ContainerStarted","Data":"80033e825e15c3c1ae7aac35c6e7d9d0ad91a4b0cb56cf9156abdf68e1c820a9"} Dec 27 07:48:05 crc kubenswrapper[4934]: I1227 07:48:05.994021 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" Dec 27 07:48:05 crc kubenswrapper[4934]: I1227 07:48:05.999351 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" Dec 27 07:48:05 crc kubenswrapper[4934]: I1227 07:48:05.999817 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" Dec 27 07:48:06 crc kubenswrapper[4934]: I1227 07:48:06.013180 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" podStartSLOduration=4.012335579 podStartE2EDuration="4.012335579s" podCreationTimestamp="2025-12-27 07:48:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:48:06.007137025 +0000 UTC m=+346.827577629" watchObservedRunningTime="2025-12-27 07:48:06.012335579 +0000 UTC m=+346.832776203" Dec 27 07:48:06 crc kubenswrapper[4934]: I1227 07:48:06.026074 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-d28xm" podStartSLOduration=1.8134709660000001 podStartE2EDuration="4.026050312s" podCreationTimestamp="2025-12-27 07:48:02 +0000 UTC" firstStartedPulling="2025-12-27 07:48:02.896232396 +0000 UTC m=+343.716672990" lastFinishedPulling="2025-12-27 07:48:05.108811742 +0000 UTC m=+345.929252336" observedRunningTime="2025-12-27 07:48:06.02186618 +0000 UTC m=+346.842306774" watchObservedRunningTime="2025-12-27 07:48:06.026050312 +0000 UTC m=+346.846490936" Dec 27 07:48:06 crc kubenswrapper[4934]: I1227 07:48:06.372348 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/964b7818-b04e-4c6d-8ad0-5a108513be73-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-2n6w2\" (UID: \"964b7818-b04e-4c6d-8ad0-5a108513be73\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" Dec 27 07:48:06 crc kubenswrapper[4934]: E1227 07:48:06.372554 4934 secret.go:188] Couldn't get secret openshift-monitoring/prometheus-operator-admission-webhook-tls: secret "prometheus-operator-admission-webhook-tls" not found Dec 27 07:48:06 crc kubenswrapper[4934]: E1227 07:48:06.372946 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/964b7818-b04e-4c6d-8ad0-5a108513be73-tls-certificates podName:964b7818-b04e-4c6d-8ad0-5a108513be73 nodeName:}" failed. No retries permitted until 2025-12-27 07:48:07.37291466 +0000 UTC m=+348.193355294 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "tls-certificates" (UniqueName: "kubernetes.io/secret/964b7818-b04e-4c6d-8ad0-5a108513be73-tls-certificates") pod "prometheus-operator-admission-webhook-f54c54754-2n6w2" (UID: "964b7818-b04e-4c6d-8ad0-5a108513be73") : secret "prometheus-operator-admission-webhook-tls" not found Dec 27 07:48:07 crc kubenswrapper[4934]: I1227 07:48:07.387328 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/964b7818-b04e-4c6d-8ad0-5a108513be73-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-2n6w2\" (UID: \"964b7818-b04e-4c6d-8ad0-5a108513be73\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" Dec 27 07:48:07 crc kubenswrapper[4934]: E1227 07:48:07.387600 4934 secret.go:188] Couldn't get secret openshift-monitoring/prometheus-operator-admission-webhook-tls: secret "prometheus-operator-admission-webhook-tls" not found Dec 27 07:48:07 crc kubenswrapper[4934]: E1227 07:48:07.387759 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/964b7818-b04e-4c6d-8ad0-5a108513be73-tls-certificates podName:964b7818-b04e-4c6d-8ad0-5a108513be73 nodeName:}" failed. No retries permitted until 2025-12-27 07:48:09.387721354 +0000 UTC m=+350.208161948 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "tls-certificates" (UniqueName: "kubernetes.io/secret/964b7818-b04e-4c6d-8ad0-5a108513be73-tls-certificates") pod "prometheus-operator-admission-webhook-f54c54754-2n6w2" (UID: "964b7818-b04e-4c6d-8ad0-5a108513be73") : secret "prometheus-operator-admission-webhook-tls" not found Dec 27 07:48:09 crc kubenswrapper[4934]: I1227 07:48:09.417459 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/964b7818-b04e-4c6d-8ad0-5a108513be73-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-2n6w2\" (UID: \"964b7818-b04e-4c6d-8ad0-5a108513be73\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" Dec 27 07:48:09 crc kubenswrapper[4934]: E1227 07:48:09.417685 4934 secret.go:188] Couldn't get secret openshift-monitoring/prometheus-operator-admission-webhook-tls: secret "prometheus-operator-admission-webhook-tls" not found Dec 27 07:48:09 crc kubenswrapper[4934]: E1227 07:48:09.417798 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/964b7818-b04e-4c6d-8ad0-5a108513be73-tls-certificates podName:964b7818-b04e-4c6d-8ad0-5a108513be73 nodeName:}" failed. No retries permitted until 2025-12-27 07:48:13.41776696 +0000 UTC m=+354.238207564 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "tls-certificates" (UniqueName: "kubernetes.io/secret/964b7818-b04e-4c6d-8ad0-5a108513be73-tls-certificates") pod "prometheus-operator-admission-webhook-f54c54754-2n6w2" (UID: "964b7818-b04e-4c6d-8ad0-5a108513be73") : secret "prometheus-operator-admission-webhook-tls" not found Dec 27 07:48:13 crc kubenswrapper[4934]: I1227 07:48:13.472139 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/964b7818-b04e-4c6d-8ad0-5a108513be73-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-2n6w2\" (UID: \"964b7818-b04e-4c6d-8ad0-5a108513be73\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" Dec 27 07:48:13 crc kubenswrapper[4934]: I1227 07:48:13.479991 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/964b7818-b04e-4c6d-8ad0-5a108513be73-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-2n6w2\" (UID: \"964b7818-b04e-4c6d-8ad0-5a108513be73\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" Dec 27 07:48:13 crc kubenswrapper[4934]: I1227 07:48:13.514303 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" Dec 27 07:48:13 crc kubenswrapper[4934]: I1227 07:48:13.922427 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2"] Dec 27 07:48:14 crc kubenswrapper[4934]: I1227 07:48:14.039417 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" event={"ID":"964b7818-b04e-4c6d-8ad0-5a108513be73","Type":"ContainerStarted","Data":"60d34794462a90794c3a9581839d95f9aae31b1f5644cb1c411cf627628599e4"} Dec 27 07:48:15 crc kubenswrapper[4934]: I1227 07:48:15.329913 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 07:48:15 crc kubenswrapper[4934]: I1227 07:48:15.330006 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.056405 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" event={"ID":"964b7818-b04e-4c6d-8ad0-5a108513be73","Type":"ContainerStarted","Data":"bf4828a80aa2309175c3712a6b88e7388112737fa373bc6e393056a56a3ad929"} Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.057140 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.064282 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.074719 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" podStartSLOduration=9.173745717 podStartE2EDuration="12.07470013s" podCreationTimestamp="2025-12-27 07:48:05 +0000 UTC" firstStartedPulling="2025-12-27 07:48:13.930570738 +0000 UTC m=+354.751011352" lastFinishedPulling="2025-12-27 07:48:16.831525161 +0000 UTC m=+357.651965765" observedRunningTime="2025-12-27 07:48:17.071719154 +0000 UTC m=+357.892159778" watchObservedRunningTime="2025-12-27 07:48:17.07470013 +0000 UTC m=+357.895140744" Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.806187 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-5js5m"] Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.807844 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-5js5m" Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.809851 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-dockercfg-fwc6j" Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.810193 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-kube-rbac-proxy-config" Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.812001 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-tls" Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.812267 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-client-ca" Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.826557 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a1306494-736d-448e-b71b-66a03e51914a-metrics-client-ca\") pod \"prometheus-operator-db54df47d-5js5m\" (UID: \"a1306494-736d-448e-b71b-66a03e51914a\") " pod="openshift-monitoring/prometheus-operator-db54df47d-5js5m" Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.826638 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xz66n\" (UniqueName: \"kubernetes.io/projected/a1306494-736d-448e-b71b-66a03e51914a-kube-api-access-xz66n\") pod \"prometheus-operator-db54df47d-5js5m\" (UID: \"a1306494-736d-448e-b71b-66a03e51914a\") " pod="openshift-monitoring/prometheus-operator-db54df47d-5js5m" Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.826841 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/a1306494-736d-448e-b71b-66a03e51914a-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-5js5m\" (UID: \"a1306494-736d-448e-b71b-66a03e51914a\") " pod="openshift-monitoring/prometheus-operator-db54df47d-5js5m" Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.826915 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/a1306494-736d-448e-b71b-66a03e51914a-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-5js5m\" (UID: \"a1306494-736d-448e-b71b-66a03e51914a\") " pod="openshift-monitoring/prometheus-operator-db54df47d-5js5m" Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.828275 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-5js5m"] Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.928690 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a1306494-736d-448e-b71b-66a03e51914a-metrics-client-ca\") pod \"prometheus-operator-db54df47d-5js5m\" (UID: \"a1306494-736d-448e-b71b-66a03e51914a\") " pod="openshift-monitoring/prometheus-operator-db54df47d-5js5m" Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.928824 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xz66n\" (UniqueName: \"kubernetes.io/projected/a1306494-736d-448e-b71b-66a03e51914a-kube-api-access-xz66n\") pod \"prometheus-operator-db54df47d-5js5m\" (UID: \"a1306494-736d-448e-b71b-66a03e51914a\") " pod="openshift-monitoring/prometheus-operator-db54df47d-5js5m" Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.928917 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/a1306494-736d-448e-b71b-66a03e51914a-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-5js5m\" (UID: \"a1306494-736d-448e-b71b-66a03e51914a\") " pod="openshift-monitoring/prometheus-operator-db54df47d-5js5m" Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.928958 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/a1306494-736d-448e-b71b-66a03e51914a-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-5js5m\" (UID: \"a1306494-736d-448e-b71b-66a03e51914a\") " pod="openshift-monitoring/prometheus-operator-db54df47d-5js5m" Dec 27 07:48:17 crc kubenswrapper[4934]: E1227 07:48:17.929231 4934 secret.go:188] Couldn't get secret openshift-monitoring/prometheus-operator-tls: secret "prometheus-operator-tls" not found Dec 27 07:48:17 crc kubenswrapper[4934]: E1227 07:48:17.929313 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a1306494-736d-448e-b71b-66a03e51914a-prometheus-operator-tls podName:a1306494-736d-448e-b71b-66a03e51914a nodeName:}" failed. No retries permitted until 2025-12-27 07:48:18.429288557 +0000 UTC m=+359.249729161 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "prometheus-operator-tls" (UniqueName: "kubernetes.io/secret/a1306494-736d-448e-b71b-66a03e51914a-prometheus-operator-tls") pod "prometheus-operator-db54df47d-5js5m" (UID: "a1306494-736d-448e-b71b-66a03e51914a") : secret "prometheus-operator-tls" not found Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.931066 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a1306494-736d-448e-b71b-66a03e51914a-metrics-client-ca\") pod \"prometheus-operator-db54df47d-5js5m\" (UID: \"a1306494-736d-448e-b71b-66a03e51914a\") " pod="openshift-monitoring/prometheus-operator-db54df47d-5js5m" Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.935913 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/a1306494-736d-448e-b71b-66a03e51914a-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-5js5m\" (UID: \"a1306494-736d-448e-b71b-66a03e51914a\") " pod="openshift-monitoring/prometheus-operator-db54df47d-5js5m" Dec 27 07:48:17 crc kubenswrapper[4934]: I1227 07:48:17.958451 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xz66n\" (UniqueName: \"kubernetes.io/projected/a1306494-736d-448e-b71b-66a03e51914a-kube-api-access-xz66n\") pod \"prometheus-operator-db54df47d-5js5m\" (UID: \"a1306494-736d-448e-b71b-66a03e51914a\") " pod="openshift-monitoring/prometheus-operator-db54df47d-5js5m" Dec 27 07:48:18 crc kubenswrapper[4934]: I1227 07:48:18.436519 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/a1306494-736d-448e-b71b-66a03e51914a-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-5js5m\" (UID: \"a1306494-736d-448e-b71b-66a03e51914a\") " pod="openshift-monitoring/prometheus-operator-db54df47d-5js5m" Dec 27 07:48:18 crc kubenswrapper[4934]: I1227 07:48:18.443851 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/a1306494-736d-448e-b71b-66a03e51914a-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-5js5m\" (UID: \"a1306494-736d-448e-b71b-66a03e51914a\") " pod="openshift-monitoring/prometheus-operator-db54df47d-5js5m" Dec 27 07:48:18 crc kubenswrapper[4934]: I1227 07:48:18.740040 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-5js5m" Dec 27 07:48:19 crc kubenswrapper[4934]: I1227 07:48:19.262823 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-5js5m"] Dec 27 07:48:19 crc kubenswrapper[4934]: W1227 07:48:19.276556 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1306494_736d_448e_b71b_66a03e51914a.slice/crio-d3974b4688284a8812bc53e4ebd8e7a85eeec724bda2b6d2a3a12b4417187cb9 WatchSource:0}: Error finding container d3974b4688284a8812bc53e4ebd8e7a85eeec724bda2b6d2a3a12b4417187cb9: Status 404 returned error can't find the container with id d3974b4688284a8812bc53e4ebd8e7a85eeec724bda2b6d2a3a12b4417187cb9 Dec 27 07:48:20 crc kubenswrapper[4934]: I1227 07:48:20.081342 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-5js5m" event={"ID":"a1306494-736d-448e-b71b-66a03e51914a","Type":"ContainerStarted","Data":"d3974b4688284a8812bc53e4ebd8e7a85eeec724bda2b6d2a3a12b4417187cb9"} Dec 27 07:48:22 crc kubenswrapper[4934]: I1227 07:48:22.099909 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-5js5m" event={"ID":"a1306494-736d-448e-b71b-66a03e51914a","Type":"ContainerStarted","Data":"910f363d1b3627771c3d3ec66e016f5518713126e182eea8f2994dbb4179ae61"} Dec 27 07:48:22 crc kubenswrapper[4934]: I1227 07:48:22.100282 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-5js5m" event={"ID":"a1306494-736d-448e-b71b-66a03e51914a","Type":"ContainerStarted","Data":"0477cd1d9203ecbe3ab0d8e1b0b0513c8240901b921fe90857c2c63baaa7ae31"} Dec 27 07:48:22 crc kubenswrapper[4934]: I1227 07:48:22.124836 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-db54df47d-5js5m" podStartSLOduration=3.189789799 podStartE2EDuration="5.124818729s" podCreationTimestamp="2025-12-27 07:48:17 +0000 UTC" firstStartedPulling="2025-12-27 07:48:19.284243979 +0000 UTC m=+360.104684573" lastFinishedPulling="2025-12-27 07:48:21.219272909 +0000 UTC m=+362.039713503" observedRunningTime="2025-12-27 07:48:22.121742451 +0000 UTC m=+362.942183045" watchObservedRunningTime="2025-12-27 07:48:22.124818729 +0000 UTC m=+362.945259323" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.182884 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-wbchw"] Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.184225 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-wbchw" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.186157 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-tls" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.186312 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-kube-rbac-proxy-config" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.186442 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-dockercfg-hvfv2" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.186895 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5"] Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.190111 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.190703 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-wbchw"] Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.192259 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-state-metrics-custom-resource-state-configmap" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.192581 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-dockercfg-zpzzc" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.192691 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-tls" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.194589 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-kube-rbac-proxy-config" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.214090 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-726lp"] Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.215004 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.216335 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/5756d91c-7c80-485e-9770-cc537f28bcbe-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-9csn5\" (UID: \"5756d91c-7c80-485e-9770-cc537f28bcbe\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.216390 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/5756d91c-7c80-485e-9770-cc537f28bcbe-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-9csn5\" (UID: \"5756d91c-7c80-485e-9770-cc537f28bcbe\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.216438 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9qx7\" (UniqueName: \"kubernetes.io/projected/5756d91c-7c80-485e-9770-cc537f28bcbe-kube-api-access-g9qx7\") pod \"kube-state-metrics-777cb5bd5d-9csn5\" (UID: \"5756d91c-7c80-485e-9770-cc537f28bcbe\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.216487 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/315af921-4327-4e80-ae8e-84b667bbb4b3-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-wbchw\" (UID: \"315af921-4327-4e80-ae8e-84b667bbb4b3\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wbchw" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.216556 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/315af921-4327-4e80-ae8e-84b667bbb4b3-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-wbchw\" (UID: \"315af921-4327-4e80-ae8e-84b667bbb4b3\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wbchw" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.216589 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/315af921-4327-4e80-ae8e-84b667bbb4b3-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-wbchw\" (UID: \"315af921-4327-4e80-ae8e-84b667bbb4b3\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wbchw" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.216617 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/5756d91c-7c80-485e-9770-cc537f28bcbe-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-9csn5\" (UID: \"5756d91c-7c80-485e-9770-cc537f28bcbe\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.216644 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/5756d91c-7c80-485e-9770-cc537f28bcbe-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-9csn5\" (UID: \"5756d91c-7c80-485e-9770-cc537f28bcbe\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.216675 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9974t\" (UniqueName: \"kubernetes.io/projected/315af921-4327-4e80-ae8e-84b667bbb4b3-kube-api-access-9974t\") pod \"openshift-state-metrics-566fddb674-wbchw\" (UID: \"315af921-4327-4e80-ae8e-84b667bbb4b3\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wbchw" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.216707 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/5756d91c-7c80-485e-9770-cc537f28bcbe-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-9csn5\" (UID: \"5756d91c-7c80-485e-9770-cc537f28bcbe\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.218707 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-tls" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.219124 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-kube-rbac-proxy-config" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.219167 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-dockercfg-x6vxz" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.224520 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5"] Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.318362 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/193fe9b7-03d1-4904-9c2e-0635f109e03a-sys\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.318419 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/193fe9b7-03d1-4904-9c2e-0635f109e03a-root\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.318493 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/193fe9b7-03d1-4904-9c2e-0635f109e03a-node-exporter-textfile\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.318519 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nghb5\" (UniqueName: \"kubernetes.io/projected/193fe9b7-03d1-4904-9c2e-0635f109e03a-kube-api-access-nghb5\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.318542 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/315af921-4327-4e80-ae8e-84b667bbb4b3-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-wbchw\" (UID: \"315af921-4327-4e80-ae8e-84b667bbb4b3\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wbchw" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.318560 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/193fe9b7-03d1-4904-9c2e-0635f109e03a-node-exporter-tls\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.318656 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/315af921-4327-4e80-ae8e-84b667bbb4b3-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-wbchw\" (UID: \"315af921-4327-4e80-ae8e-84b667bbb4b3\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wbchw" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.318715 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/5756d91c-7c80-485e-9770-cc537f28bcbe-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-9csn5\" (UID: \"5756d91c-7c80-485e-9770-cc537f28bcbe\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.318769 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/5756d91c-7c80-485e-9770-cc537f28bcbe-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-9csn5\" (UID: \"5756d91c-7c80-485e-9770-cc537f28bcbe\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.318818 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9974t\" (UniqueName: \"kubernetes.io/projected/315af921-4327-4e80-ae8e-84b667bbb4b3-kube-api-access-9974t\") pod \"openshift-state-metrics-566fddb674-wbchw\" (UID: \"315af921-4327-4e80-ae8e-84b667bbb4b3\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wbchw" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.318869 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/5756d91c-7c80-485e-9770-cc537f28bcbe-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-9csn5\" (UID: \"5756d91c-7c80-485e-9770-cc537f28bcbe\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" Dec 27 07:48:24 crc kubenswrapper[4934]: E1227 07:48:24.318891 4934 secret.go:188] Couldn't get secret openshift-monitoring/kube-state-metrics-tls: secret "kube-state-metrics-tls" not found Dec 27 07:48:24 crc kubenswrapper[4934]: E1227 07:48:24.318960 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5756d91c-7c80-485e-9770-cc537f28bcbe-kube-state-metrics-tls podName:5756d91c-7c80-485e-9770-cc537f28bcbe nodeName:}" failed. No retries permitted until 2025-12-27 07:48:24.818940699 +0000 UTC m=+365.639381293 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-state-metrics-tls" (UniqueName: "kubernetes.io/secret/5756d91c-7c80-485e-9770-cc537f28bcbe-kube-state-metrics-tls") pod "kube-state-metrics-777cb5bd5d-9csn5" (UID: "5756d91c-7c80-485e-9770-cc537f28bcbe") : secret "kube-state-metrics-tls" not found Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.318896 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/5756d91c-7c80-485e-9770-cc537f28bcbe-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-9csn5\" (UID: \"5756d91c-7c80-485e-9770-cc537f28bcbe\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.319004 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/5756d91c-7c80-485e-9770-cc537f28bcbe-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-9csn5\" (UID: \"5756d91c-7c80-485e-9770-cc537f28bcbe\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.319026 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9qx7\" (UniqueName: \"kubernetes.io/projected/5756d91c-7c80-485e-9770-cc537f28bcbe-kube-api-access-g9qx7\") pod \"kube-state-metrics-777cb5bd5d-9csn5\" (UID: \"5756d91c-7c80-485e-9770-cc537f28bcbe\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.319101 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/193fe9b7-03d1-4904-9c2e-0635f109e03a-node-exporter-wtmp\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.319119 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/193fe9b7-03d1-4904-9c2e-0635f109e03a-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.319143 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/193fe9b7-03d1-4904-9c2e-0635f109e03a-metrics-client-ca\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.319168 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/315af921-4327-4e80-ae8e-84b667bbb4b3-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-wbchw\" (UID: \"315af921-4327-4e80-ae8e-84b667bbb4b3\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wbchw" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.319549 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/5756d91c-7c80-485e-9770-cc537f28bcbe-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-9csn5\" (UID: \"5756d91c-7c80-485e-9770-cc537f28bcbe\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.319620 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/315af921-4327-4e80-ae8e-84b667bbb4b3-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-wbchw\" (UID: \"315af921-4327-4e80-ae8e-84b667bbb4b3\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wbchw" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.319677 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/5756d91c-7c80-485e-9770-cc537f28bcbe-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-9csn5\" (UID: \"5756d91c-7c80-485e-9770-cc537f28bcbe\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" Dec 27 07:48:24 crc kubenswrapper[4934]: E1227 07:48:24.319845 4934 secret.go:188] Couldn't get secret openshift-monitoring/openshift-state-metrics-tls: secret "openshift-state-metrics-tls" not found Dec 27 07:48:24 crc kubenswrapper[4934]: E1227 07:48:24.319875 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/315af921-4327-4e80-ae8e-84b667bbb4b3-openshift-state-metrics-tls podName:315af921-4327-4e80-ae8e-84b667bbb4b3 nodeName:}" failed. No retries permitted until 2025-12-27 07:48:24.819865719 +0000 UTC m=+365.640306313 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openshift-state-metrics-tls" (UniqueName: "kubernetes.io/secret/315af921-4327-4e80-ae8e-84b667bbb4b3-openshift-state-metrics-tls") pod "openshift-state-metrics-566fddb674-wbchw" (UID: "315af921-4327-4e80-ae8e-84b667bbb4b3") : secret "openshift-state-metrics-tls" not found Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.320250 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/5756d91c-7c80-485e-9770-cc537f28bcbe-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-9csn5\" (UID: \"5756d91c-7c80-485e-9770-cc537f28bcbe\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.325054 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/315af921-4327-4e80-ae8e-84b667bbb4b3-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-wbchw\" (UID: \"315af921-4327-4e80-ae8e-84b667bbb4b3\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wbchw" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.325180 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/5756d91c-7c80-485e-9770-cc537f28bcbe-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-9csn5\" (UID: \"5756d91c-7c80-485e-9770-cc537f28bcbe\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.338352 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9974t\" (UniqueName: \"kubernetes.io/projected/315af921-4327-4e80-ae8e-84b667bbb4b3-kube-api-access-9974t\") pod \"openshift-state-metrics-566fddb674-wbchw\" (UID: \"315af921-4327-4e80-ae8e-84b667bbb4b3\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wbchw" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.338661 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9qx7\" (UniqueName: \"kubernetes.io/projected/5756d91c-7c80-485e-9770-cc537f28bcbe-kube-api-access-g9qx7\") pod \"kube-state-metrics-777cb5bd5d-9csn5\" (UID: \"5756d91c-7c80-485e-9770-cc537f28bcbe\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.420666 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/193fe9b7-03d1-4904-9c2e-0635f109e03a-node-exporter-textfile\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.420720 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nghb5\" (UniqueName: \"kubernetes.io/projected/193fe9b7-03d1-4904-9c2e-0635f109e03a-kube-api-access-nghb5\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.420741 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/193fe9b7-03d1-4904-9c2e-0635f109e03a-node-exporter-tls\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.420815 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/193fe9b7-03d1-4904-9c2e-0635f109e03a-node-exporter-wtmp\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.420830 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/193fe9b7-03d1-4904-9c2e-0635f109e03a-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.420853 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/193fe9b7-03d1-4904-9c2e-0635f109e03a-metrics-client-ca\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.420872 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/193fe9b7-03d1-4904-9c2e-0635f109e03a-sys\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.420887 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/193fe9b7-03d1-4904-9c2e-0635f109e03a-root\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.420952 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/193fe9b7-03d1-4904-9c2e-0635f109e03a-root\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.421280 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/193fe9b7-03d1-4904-9c2e-0635f109e03a-node-exporter-textfile\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.421988 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/193fe9b7-03d1-4904-9c2e-0635f109e03a-sys\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.422047 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/193fe9b7-03d1-4904-9c2e-0635f109e03a-node-exporter-wtmp\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.422452 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/193fe9b7-03d1-4904-9c2e-0635f109e03a-metrics-client-ca\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.424305 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/193fe9b7-03d1-4904-9c2e-0635f109e03a-node-exporter-tls\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.424765 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/193fe9b7-03d1-4904-9c2e-0635f109e03a-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.457722 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nghb5\" (UniqueName: \"kubernetes.io/projected/193fe9b7-03d1-4904-9c2e-0635f109e03a-kube-api-access-nghb5\") pod \"node-exporter-726lp\" (UID: \"193fe9b7-03d1-4904-9c2e-0635f109e03a\") " pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.529254 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-726lp" Dec 27 07:48:24 crc kubenswrapper[4934]: W1227 07:48:24.544401 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod193fe9b7_03d1_4904_9c2e_0635f109e03a.slice/crio-c9d5b6803e132071b048e57681672642da55e025cc45882e2fa5a315460dda01 WatchSource:0}: Error finding container c9d5b6803e132071b048e57681672642da55e025cc45882e2fa5a315460dda01: Status 404 returned error can't find the container with id c9d5b6803e132071b048e57681672642da55e025cc45882e2fa5a315460dda01 Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.825978 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/315af921-4327-4e80-ae8e-84b667bbb4b3-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-wbchw\" (UID: \"315af921-4327-4e80-ae8e-84b667bbb4b3\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wbchw" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.826059 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/5756d91c-7c80-485e-9770-cc537f28bcbe-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-9csn5\" (UID: \"5756d91c-7c80-485e-9770-cc537f28bcbe\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.830887 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/5756d91c-7c80-485e-9770-cc537f28bcbe-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-9csn5\" (UID: \"5756d91c-7c80-485e-9770-cc537f28bcbe\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" Dec 27 07:48:24 crc kubenswrapper[4934]: I1227 07:48:24.832222 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/315af921-4327-4e80-ae8e-84b667bbb4b3-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-wbchw\" (UID: \"315af921-4327-4e80-ae8e-84b667bbb4b3\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-wbchw" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.106520 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-wbchw" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.118852 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.119378 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-726lp" event={"ID":"193fe9b7-03d1-4904-9c2e-0635f109e03a","Type":"ContainerStarted","Data":"c9d5b6803e132071b048e57681672642da55e025cc45882e2fa5a315460dda01"} Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.274128 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.282604 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.286247 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-generated" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.286291 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-web" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.286568 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-web-config" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.286574 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls-assets-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.286614 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-metric" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.286964 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.287377 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.287529 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-dockercfg-prtbc" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.298286 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"alertmanager-trusted-ca-bundle" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.313857 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.335493 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtjvh\" (UniqueName: \"kubernetes.io/projected/17fe170d-1e03-46fb-a0ea-3b209cb114ff-kube-api-access-qtjvh\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.335539 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/17fe170d-1e03-46fb-a0ea-3b209cb114ff-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.335562 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/17fe170d-1e03-46fb-a0ea-3b209cb114ff-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.335582 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/17fe170d-1e03-46fb-a0ea-3b209cb114ff-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.335599 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/17fe170d-1e03-46fb-a0ea-3b209cb114ff-tls-assets\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.335639 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/17fe170d-1e03-46fb-a0ea-3b209cb114ff-config-out\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.335656 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/17fe170d-1e03-46fb-a0ea-3b209cb114ff-web-config\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.335692 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/17fe170d-1e03-46fb-a0ea-3b209cb114ff-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.335714 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/17fe170d-1e03-46fb-a0ea-3b209cb114ff-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.335732 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/17fe170d-1e03-46fb-a0ea-3b209cb114ff-config-volume\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.335750 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/17fe170d-1e03-46fb-a0ea-3b209cb114ff-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.335769 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/17fe170d-1e03-46fb-a0ea-3b209cb114ff-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.436869 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/17fe170d-1e03-46fb-a0ea-3b209cb114ff-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.436955 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/17fe170d-1e03-46fb-a0ea-3b209cb114ff-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.436988 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/17fe170d-1e03-46fb-a0ea-3b209cb114ff-config-volume\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.437042 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/17fe170d-1e03-46fb-a0ea-3b209cb114ff-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.437495 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/17fe170d-1e03-46fb-a0ea-3b209cb114ff-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.438687 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/17fe170d-1e03-46fb-a0ea-3b209cb114ff-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.439376 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtjvh\" (UniqueName: \"kubernetes.io/projected/17fe170d-1e03-46fb-a0ea-3b209cb114ff-kube-api-access-qtjvh\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.439451 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/17fe170d-1e03-46fb-a0ea-3b209cb114ff-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.439613 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/17fe170d-1e03-46fb-a0ea-3b209cb114ff-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.439641 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/17fe170d-1e03-46fb-a0ea-3b209cb114ff-tls-assets\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.439661 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/17fe170d-1e03-46fb-a0ea-3b209cb114ff-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.439698 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/17fe170d-1e03-46fb-a0ea-3b209cb114ff-config-out\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.440107 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/17fe170d-1e03-46fb-a0ea-3b209cb114ff-web-config\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.444645 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/17fe170d-1e03-46fb-a0ea-3b209cb114ff-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.444663 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/17fe170d-1e03-46fb-a0ea-3b209cb114ff-tls-assets\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.449296 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/17fe170d-1e03-46fb-a0ea-3b209cb114ff-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.449873 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/17fe170d-1e03-46fb-a0ea-3b209cb114ff-config-volume\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.460523 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/17fe170d-1e03-46fb-a0ea-3b209cb114ff-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.460689 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/17fe170d-1e03-46fb-a0ea-3b209cb114ff-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.461557 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/17fe170d-1e03-46fb-a0ea-3b209cb114ff-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.465887 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtjvh\" (UniqueName: \"kubernetes.io/projected/17fe170d-1e03-46fb-a0ea-3b209cb114ff-kube-api-access-qtjvh\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.466881 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/17fe170d-1e03-46fb-a0ea-3b209cb114ff-web-config\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.469468 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/17fe170d-1e03-46fb-a0ea-3b209cb114ff-config-out\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.477846 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/17fe170d-1e03-46fb-a0ea-3b209cb114ff-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"17fe170d-1e03-46fb-a0ea-3b209cb114ff\") " pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.612308 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.620431 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5"] Dec 27 07:48:25 crc kubenswrapper[4934]: I1227 07:48:25.770305 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-wbchw"] Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.049306 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.126142 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" event={"ID":"5756d91c-7c80-485e-9770-cc537f28bcbe","Type":"ContainerStarted","Data":"66c1eaed0139c41c446b054d0a135cc6419d02c864bcb5c287f7eeeb6fcc7aec"} Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.128333 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/thanos-querier-8467887986-svb7h"] Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.133647 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.137416 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-rules" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.137442 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-dockercfg-xcxgh" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.137642 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-metrics" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.138019 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-grpc-tls-u4javjlce0up" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.139590 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-web" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.139712 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.139723 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-tls" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.155795 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-8467887986-svb7h"] Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.255331 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/a07fdfc4-584a-4a39-b87f-ee78f9cac687-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.255424 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/a07fdfc4-584a-4a39-b87f-ee78f9cac687-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.255466 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a07fdfc4-584a-4a39-b87f-ee78f9cac687-metrics-client-ca\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.255492 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/a07fdfc4-584a-4a39-b87f-ee78f9cac687-secret-grpc-tls\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.255602 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/a07fdfc4-584a-4a39-b87f-ee78f9cac687-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.255664 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/a07fdfc4-584a-4a39-b87f-ee78f9cac687-secret-thanos-querier-tls\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.255748 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8cwk\" (UniqueName: \"kubernetes.io/projected/a07fdfc4-584a-4a39-b87f-ee78f9cac687-kube-api-access-n8cwk\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.255777 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/a07fdfc4-584a-4a39-b87f-ee78f9cac687-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.357586 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/a07fdfc4-584a-4a39-b87f-ee78f9cac687-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.357717 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/a07fdfc4-584a-4a39-b87f-ee78f9cac687-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.357812 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a07fdfc4-584a-4a39-b87f-ee78f9cac687-metrics-client-ca\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.357885 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/a07fdfc4-584a-4a39-b87f-ee78f9cac687-secret-grpc-tls\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.357954 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/a07fdfc4-584a-4a39-b87f-ee78f9cac687-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.358022 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/a07fdfc4-584a-4a39-b87f-ee78f9cac687-secret-thanos-querier-tls\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.358140 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8cwk\" (UniqueName: \"kubernetes.io/projected/a07fdfc4-584a-4a39-b87f-ee78f9cac687-kube-api-access-n8cwk\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.358194 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/a07fdfc4-584a-4a39-b87f-ee78f9cac687-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.360554 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a07fdfc4-584a-4a39-b87f-ee78f9cac687-metrics-client-ca\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.363227 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/a07fdfc4-584a-4a39-b87f-ee78f9cac687-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.364816 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/a07fdfc4-584a-4a39-b87f-ee78f9cac687-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.366030 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/a07fdfc4-584a-4a39-b87f-ee78f9cac687-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.369300 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/a07fdfc4-584a-4a39-b87f-ee78f9cac687-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.370768 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/a07fdfc4-584a-4a39-b87f-ee78f9cac687-secret-thanos-querier-tls\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.371582 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/a07fdfc4-584a-4a39-b87f-ee78f9cac687-secret-grpc-tls\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.378361 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8cwk\" (UniqueName: \"kubernetes.io/projected/a07fdfc4-584a-4a39-b87f-ee78f9cac687-kube-api-access-n8cwk\") pod \"thanos-querier-8467887986-svb7h\" (UID: \"a07fdfc4-584a-4a39-b87f-ee78f9cac687\") " pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.450228 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:26 crc kubenswrapper[4934]: I1227 07:48:26.913753 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-8467887986-svb7h"] Dec 27 07:48:26 crc kubenswrapper[4934]: W1227 07:48:26.922996 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda07fdfc4_584a_4a39_b87f_ee78f9cac687.slice/crio-857571738d8d168b98f8785d35325d340058f19bb73fd20cc72f1e9ae8ad019d WatchSource:0}: Error finding container 857571738d8d168b98f8785d35325d340058f19bb73fd20cc72f1e9ae8ad019d: Status 404 returned error can't find the container with id 857571738d8d168b98f8785d35325d340058f19bb73fd20cc72f1e9ae8ad019d Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.133717 4934 generic.go:334] "Generic (PLEG): container finished" podID="193fe9b7-03d1-4904-9c2e-0635f109e03a" containerID="273c7230e4a3be39090a5ad3c0d7ff6adef7423de7cd35f22cdbfc401f66e36a" exitCode=0 Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.133793 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-726lp" event={"ID":"193fe9b7-03d1-4904-9c2e-0635f109e03a","Type":"ContainerDied","Data":"273c7230e4a3be39090a5ad3c0d7ff6adef7423de7cd35f22cdbfc401f66e36a"} Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.138037 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-wbchw" event={"ID":"315af921-4327-4e80-ae8e-84b667bbb4b3","Type":"ContainerStarted","Data":"fdfde136969e4efe35ad79ce20529cffea20b34b49f9b373108a8004fad70750"} Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.138104 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-wbchw" event={"ID":"315af921-4327-4e80-ae8e-84b667bbb4b3","Type":"ContainerStarted","Data":"394f8862e031a4486072fc70b7aad626a54eb91df6182e727a9cc69526d23963"} Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.138118 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-wbchw" event={"ID":"315af921-4327-4e80-ae8e-84b667bbb4b3","Type":"ContainerStarted","Data":"dcfc3998722ed6b7051232e7b0e1287b089b3c498ea77ad82bacd3f311fc38ae"} Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.138936 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"17fe170d-1e03-46fb-a0ea-3b209cb114ff","Type":"ContainerStarted","Data":"1944b8b293c5215c72a09deeab1088e723ee81b38798cbdc52af589c83a06896"} Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.139802 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-8467887986-svb7h" event={"ID":"a07fdfc4-584a-4a39-b87f-ee78f9cac687","Type":"ContainerStarted","Data":"857571738d8d168b98f8785d35325d340058f19bb73fd20cc72f1e9ae8ad019d"} Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.161258 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" podUID="c2f61301-d046-4651-b76c-9939b18fb792" containerName="oauth-openshift" containerID="cri-o://e7fc2fe75e37d007bed71c48bbf23cadc6115308f46af37e1bcd5966c0e7037c" gracePeriod=15 Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.823342 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.854870 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-fd9565bb5-mkr88"] Dec 27 07:48:27 crc kubenswrapper[4934]: E1227 07:48:27.855112 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2f61301-d046-4651-b76c-9939b18fb792" containerName="oauth-openshift" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.855126 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2f61301-d046-4651-b76c-9939b18fb792" containerName="oauth-openshift" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.855243 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2f61301-d046-4651-b76c-9939b18fb792" containerName="oauth-openshift" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.855674 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.871361 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-fd9565bb5-mkr88"] Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.884224 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-session\") pod \"c2f61301-d046-4651-b76c-9939b18fb792\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.884560 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-ocp-branding-template\") pod \"c2f61301-d046-4651-b76c-9939b18fb792\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.884599 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-serving-cert\") pod \"c2f61301-d046-4651-b76c-9939b18fb792\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.884624 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-cliconfig\") pod \"c2f61301-d046-4651-b76c-9939b18fb792\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.884641 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpk5n\" (UniqueName: \"kubernetes.io/projected/c2f61301-d046-4651-b76c-9939b18fb792-kube-api-access-jpk5n\") pod \"c2f61301-d046-4651-b76c-9939b18fb792\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.884679 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-router-certs\") pod \"c2f61301-d046-4651-b76c-9939b18fb792\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.884725 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-template-provider-selection\") pod \"c2f61301-d046-4651-b76c-9939b18fb792\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.884747 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-trusted-ca-bundle\") pod \"c2f61301-d046-4651-b76c-9939b18fb792\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.884768 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-template-login\") pod \"c2f61301-d046-4651-b76c-9939b18fb792\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.884788 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-audit-policies\") pod \"c2f61301-d046-4651-b76c-9939b18fb792\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.884806 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-service-ca\") pod \"c2f61301-d046-4651-b76c-9939b18fb792\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.884827 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-idp-0-file-data\") pod \"c2f61301-d046-4651-b76c-9939b18fb792\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.884890 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c2f61301-d046-4651-b76c-9939b18fb792-audit-dir\") pod \"c2f61301-d046-4651-b76c-9939b18fb792\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.884924 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-template-error\") pod \"c2f61301-d046-4651-b76c-9939b18fb792\" (UID: \"c2f61301-d046-4651-b76c-9939b18fb792\") " Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.885060 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/772a3e80-411a-4c15-9e1c-b702392e13af-audit-policies\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.885097 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.885731 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.885765 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-user-template-login\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.885780 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.885805 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/772a3e80-411a-4c15-9e1c-b702392e13af-audit-dir\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.885823 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-system-session\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.885848 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-system-router-certs\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.885881 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5q9p\" (UniqueName: \"kubernetes.io/projected/772a3e80-411a-4c15-9e1c-b702392e13af-kube-api-access-v5q9p\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.885916 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-system-cliconfig\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.885931 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.885951 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-system-serving-cert\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.885972 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-system-service-ca\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.885988 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-user-template-error\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.891086 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "c2f61301-d046-4651-b76c-9939b18fb792" (UID: "c2f61301-d046-4651-b76c-9939b18fb792"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.892686 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "c2f61301-d046-4651-b76c-9939b18fb792" (UID: "c2f61301-d046-4651-b76c-9939b18fb792"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.893939 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "c2f61301-d046-4651-b76c-9939b18fb792" (UID: "c2f61301-d046-4651-b76c-9939b18fb792"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.894131 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c2f61301-d046-4651-b76c-9939b18fb792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "c2f61301-d046-4651-b76c-9939b18fb792" (UID: "c2f61301-d046-4651-b76c-9939b18fb792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.894603 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "c2f61301-d046-4651-b76c-9939b18fb792" (UID: "c2f61301-d046-4651-b76c-9939b18fb792"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.894877 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "c2f61301-d046-4651-b76c-9939b18fb792" (UID: "c2f61301-d046-4651-b76c-9939b18fb792"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.895936 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2f61301-d046-4651-b76c-9939b18fb792-kube-api-access-jpk5n" (OuterVolumeSpecName: "kube-api-access-jpk5n") pod "c2f61301-d046-4651-b76c-9939b18fb792" (UID: "c2f61301-d046-4651-b76c-9939b18fb792"). InnerVolumeSpecName "kube-api-access-jpk5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.898064 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "c2f61301-d046-4651-b76c-9939b18fb792" (UID: "c2f61301-d046-4651-b76c-9939b18fb792"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.899964 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "c2f61301-d046-4651-b76c-9939b18fb792" (UID: "c2f61301-d046-4651-b76c-9939b18fb792"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.900140 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "c2f61301-d046-4651-b76c-9939b18fb792" (UID: "c2f61301-d046-4651-b76c-9939b18fb792"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.900500 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "c2f61301-d046-4651-b76c-9939b18fb792" (UID: "c2f61301-d046-4651-b76c-9939b18fb792"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.900769 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "c2f61301-d046-4651-b76c-9939b18fb792" (UID: "c2f61301-d046-4651-b76c-9939b18fb792"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.900857 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "c2f61301-d046-4651-b76c-9939b18fb792" (UID: "c2f61301-d046-4651-b76c-9939b18fb792"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.904770 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "c2f61301-d046-4651-b76c-9939b18fb792" (UID: "c2f61301-d046-4651-b76c-9939b18fb792"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987028 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-user-template-login\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987087 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987142 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/772a3e80-411a-4c15-9e1c-b702392e13af-audit-dir\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987174 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-system-session\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987210 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-system-router-certs\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987236 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5q9p\" (UniqueName: \"kubernetes.io/projected/772a3e80-411a-4c15-9e1c-b702392e13af-kube-api-access-v5q9p\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987289 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987320 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-system-cliconfig\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987353 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-system-serving-cert\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987383 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-system-service-ca\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987409 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-user-template-error\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987447 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/772a3e80-411a-4c15-9e1c-b702392e13af-audit-policies\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987488 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987524 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987579 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987595 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987609 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987622 4934 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987636 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987653 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987666 4934 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c2f61301-d046-4651-b76c-9939b18fb792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987679 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987692 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987703 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987716 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987728 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987741 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpk5n\" (UniqueName: \"kubernetes.io/projected/c2f61301-d046-4651-b76c-9939b18fb792-kube-api-access-jpk5n\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.987754 4934 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c2f61301-d046-4651-b76c-9939b18fb792-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.988743 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.990658 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-system-service-ca\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.991266 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-system-cliconfig\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.991301 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-user-template-login\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.992101 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.993628 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-user-template-error\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.994224 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/772a3e80-411a-4c15-9e1c-b702392e13af-audit-policies\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.994434 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-system-session\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.995480 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-system-router-certs\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.997344 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/772a3e80-411a-4c15-9e1c-b702392e13af-audit-dir\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.997547 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:27 crc kubenswrapper[4934]: I1227 07:48:27.998775 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-system-serving-cert\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:28 crc kubenswrapper[4934]: I1227 07:48:28.001610 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/772a3e80-411a-4c15-9e1c-b702392e13af-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:28 crc kubenswrapper[4934]: I1227 07:48:28.016548 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5q9p\" (UniqueName: \"kubernetes.io/projected/772a3e80-411a-4c15-9e1c-b702392e13af-kube-api-access-v5q9p\") pod \"oauth-openshift-fd9565bb5-mkr88\" (UID: \"772a3e80-411a-4c15-9e1c-b702392e13af\") " pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:28 crc kubenswrapper[4934]: I1227 07:48:28.148504 4934 generic.go:334] "Generic (PLEG): container finished" podID="c2f61301-d046-4651-b76c-9939b18fb792" containerID="e7fc2fe75e37d007bed71c48bbf23cadc6115308f46af37e1bcd5966c0e7037c" exitCode=0 Dec 27 07:48:28 crc kubenswrapper[4934]: I1227 07:48:28.148580 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" event={"ID":"c2f61301-d046-4651-b76c-9939b18fb792","Type":"ContainerDied","Data":"e7fc2fe75e37d007bed71c48bbf23cadc6115308f46af37e1bcd5966c0e7037c"} Dec 27 07:48:28 crc kubenswrapper[4934]: I1227 07:48:28.148611 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" event={"ID":"c2f61301-d046-4651-b76c-9939b18fb792","Type":"ContainerDied","Data":"fed642fe85c2daef00a5d53c79d9c70728e906b17d0c52170df1ec440f20b6ba"} Dec 27 07:48:28 crc kubenswrapper[4934]: I1227 07:48:28.148631 4934 scope.go:117] "RemoveContainer" containerID="e7fc2fe75e37d007bed71c48bbf23cadc6115308f46af37e1bcd5966c0e7037c" Dec 27 07:48:28 crc kubenswrapper[4934]: I1227 07:48:28.148751 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-j5cpz" Dec 27 07:48:28 crc kubenswrapper[4934]: I1227 07:48:28.154184 4934 generic.go:334] "Generic (PLEG): container finished" podID="17fe170d-1e03-46fb-a0ea-3b209cb114ff" containerID="2339b5f2b6a3398653603ae980603d36e116a5a8cf8ce5a8b183523d605798d8" exitCode=0 Dec 27 07:48:28 crc kubenswrapper[4934]: I1227 07:48:28.154619 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"17fe170d-1e03-46fb-a0ea-3b209cb114ff","Type":"ContainerDied","Data":"2339b5f2b6a3398653603ae980603d36e116a5a8cf8ce5a8b183523d605798d8"} Dec 27 07:48:28 crc kubenswrapper[4934]: I1227 07:48:28.157304 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" event={"ID":"5756d91c-7c80-485e-9770-cc537f28bcbe","Type":"ContainerStarted","Data":"576d7b6cf234854d8b3d4aa1142e05f13293c4c5026e80fa5c6409dcedeb1f91"} Dec 27 07:48:28 crc kubenswrapper[4934]: I1227 07:48:28.162628 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-726lp" event={"ID":"193fe9b7-03d1-4904-9c2e-0635f109e03a","Type":"ContainerStarted","Data":"ca9ec59a25c65681635bb0c231009cebf290e68b26d766e887879ba1484233f7"} Dec 27 07:48:28 crc kubenswrapper[4934]: I1227 07:48:28.204074 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-j5cpz"] Dec 27 07:48:28 crc kubenswrapper[4934]: I1227 07:48:28.208390 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-j5cpz"] Dec 27 07:48:28 crc kubenswrapper[4934]: I1227 07:48:28.213209 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:28 crc kubenswrapper[4934]: I1227 07:48:28.472982 4934 scope.go:117] "RemoveContainer" containerID="e7fc2fe75e37d007bed71c48bbf23cadc6115308f46af37e1bcd5966c0e7037c" Dec 27 07:48:28 crc kubenswrapper[4934]: E1227 07:48:28.473566 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7fc2fe75e37d007bed71c48bbf23cadc6115308f46af37e1bcd5966c0e7037c\": container with ID starting with e7fc2fe75e37d007bed71c48bbf23cadc6115308f46af37e1bcd5966c0e7037c not found: ID does not exist" containerID="e7fc2fe75e37d007bed71c48bbf23cadc6115308f46af37e1bcd5966c0e7037c" Dec 27 07:48:28 crc kubenswrapper[4934]: I1227 07:48:28.473612 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7fc2fe75e37d007bed71c48bbf23cadc6115308f46af37e1bcd5966c0e7037c"} err="failed to get container status \"e7fc2fe75e37d007bed71c48bbf23cadc6115308f46af37e1bcd5966c0e7037c\": rpc error: code = NotFound desc = could not find container \"e7fc2fe75e37d007bed71c48bbf23cadc6115308f46af37e1bcd5966c0e7037c\": container with ID starting with e7fc2fe75e37d007bed71c48bbf23cadc6115308f46af37e1bcd5966c0e7037c not found: ID does not exist" Dec 27 07:48:28 crc kubenswrapper[4934]: I1227 07:48:28.905960 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-fd9565bb5-mkr88"] Dec 27 07:48:28 crc kubenswrapper[4934]: I1227 07:48:28.972916 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-c8c89d49c-b97s8"] Dec 27 07:48:28 crc kubenswrapper[4934]: I1227 07:48:28.973580 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.035298 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-c8c89d49c-b97s8"] Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.109350 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2f5h\" (UniqueName: \"kubernetes.io/projected/6aa1cec9-a0f2-46da-ae87-b94936a1360b-kube-api-access-d2f5h\") pod \"console-c8c89d49c-b97s8\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.109396 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-service-ca\") pod \"console-c8c89d49c-b97s8\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.109434 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-oauth-serving-cert\") pod \"console-c8c89d49c-b97s8\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.109525 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-console-config\") pod \"console-c8c89d49c-b97s8\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.109555 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6aa1cec9-a0f2-46da-ae87-b94936a1360b-console-oauth-config\") pod \"console-c8c89d49c-b97s8\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.109599 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-trusted-ca-bundle\") pod \"console-c8c89d49c-b97s8\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.109637 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6aa1cec9-a0f2-46da-ae87-b94936a1360b-console-serving-cert\") pod \"console-c8c89d49c-b97s8\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.169358 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-wbchw" event={"ID":"315af921-4327-4e80-ae8e-84b667bbb4b3","Type":"ContainerStarted","Data":"003047ea680695e29bd44d6d1eb65a2a425f5625fa2c43e4a78a4d34fbb751fa"} Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.174196 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" event={"ID":"5756d91c-7c80-485e-9770-cc537f28bcbe","Type":"ContainerStarted","Data":"996596cad65765c4271b546f07039eb9786244970bafcbc6b377279e4ce09477"} Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.174237 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" event={"ID":"5756d91c-7c80-485e-9770-cc537f28bcbe","Type":"ContainerStarted","Data":"e6fcc1e7f8012cc6e1d5ea05c09b4318b304da19ed9f254b2dbbc2fcab3fb264"} Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.176836 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-726lp" event={"ID":"193fe9b7-03d1-4904-9c2e-0635f109e03a","Type":"ContainerStarted","Data":"3d7102de03333a8a108b02970b2f2f754ea6554374d268b197b01b6ecfb0315d"} Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.192117 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/openshift-state-metrics-566fddb674-wbchw" podStartSLOduration=3.425628473 podStartE2EDuration="5.192085861s" podCreationTimestamp="2025-12-27 07:48:24 +0000 UTC" firstStartedPulling="2025-12-27 07:48:26.783533238 +0000 UTC m=+367.603973832" lastFinishedPulling="2025-12-27 07:48:28.549990616 +0000 UTC m=+369.370431220" observedRunningTime="2025-12-27 07:48:29.190183029 +0000 UTC m=+370.010623623" watchObservedRunningTime="2025-12-27 07:48:29.192085861 +0000 UTC m=+370.012526465" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.210470 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-726lp" podStartSLOduration=3.3299243609999998 podStartE2EDuration="5.210450026s" podCreationTimestamp="2025-12-27 07:48:24 +0000 UTC" firstStartedPulling="2025-12-27 07:48:24.546011262 +0000 UTC m=+365.366451856" lastFinishedPulling="2025-12-27 07:48:26.426536927 +0000 UTC m=+367.246977521" observedRunningTime="2025-12-27 07:48:29.206745664 +0000 UTC m=+370.027186258" watchObservedRunningTime="2025-12-27 07:48:29.210450026 +0000 UTC m=+370.030890620" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.211230 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-oauth-serving-cert\") pod \"console-c8c89d49c-b97s8\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.211362 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-console-config\") pod \"console-c8c89d49c-b97s8\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.211394 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6aa1cec9-a0f2-46da-ae87-b94936a1360b-console-oauth-config\") pod \"console-c8c89d49c-b97s8\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.211470 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-trusted-ca-bundle\") pod \"console-c8c89d49c-b97s8\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.211513 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6aa1cec9-a0f2-46da-ae87-b94936a1360b-console-serving-cert\") pod \"console-c8c89d49c-b97s8\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.211591 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2f5h\" (UniqueName: \"kubernetes.io/projected/6aa1cec9-a0f2-46da-ae87-b94936a1360b-kube-api-access-d2f5h\") pod \"console-c8c89d49c-b97s8\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.211639 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-service-ca\") pod \"console-c8c89d49c-b97s8\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.212843 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-service-ca\") pod \"console-c8c89d49c-b97s8\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.213240 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-oauth-serving-cert\") pod \"console-c8c89d49c-b97s8\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.213809 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-console-config\") pod \"console-c8c89d49c-b97s8\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.215288 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-trusted-ca-bundle\") pod \"console-c8c89d49c-b97s8\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.219672 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6aa1cec9-a0f2-46da-ae87-b94936a1360b-console-serving-cert\") pod \"console-c8c89d49c-b97s8\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.223132 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6aa1cec9-a0f2-46da-ae87-b94936a1360b-console-oauth-config\") pod \"console-c8c89d49c-b97s8\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.228517 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-9csn5" podStartSLOduration=3.054789986 podStartE2EDuration="5.228505565s" podCreationTimestamp="2025-12-27 07:48:24 +0000 UTC" firstStartedPulling="2025-12-27 07:48:25.636746451 +0000 UTC m=+366.457187045" lastFinishedPulling="2025-12-27 07:48:27.81046203 +0000 UTC m=+368.630902624" observedRunningTime="2025-12-27 07:48:29.225409696 +0000 UTC m=+370.045850290" watchObservedRunningTime="2025-12-27 07:48:29.228505565 +0000 UTC m=+370.048946159" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.229763 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2f5h\" (UniqueName: \"kubernetes.io/projected/6aa1cec9-a0f2-46da-ae87-b94936a1360b-kube-api-access-d2f5h\") pod \"console-c8c89d49c-b97s8\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.290394 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.371714 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/metrics-server-76d5d9b996-jzgqz"] Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.372368 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.376575 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kubelet-serving-ca-bundle" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.382589 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-76d5d9b996-jzgqz"] Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.382856 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-tls" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.385353 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-server-audit-profiles" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.385544 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-dockercfg-f7vpq" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.385655 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-client-certs" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.386188 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-1s16adgkvuime" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.414015 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/169cb9e6-0d57-425d-a51d-a3b0e1556db6-secret-metrics-client-certs\") pod \"metrics-server-76d5d9b996-jzgqz\" (UID: \"169cb9e6-0d57-425d-a51d-a3b0e1556db6\") " pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.414163 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/169cb9e6-0d57-425d-a51d-a3b0e1556db6-audit-log\") pod \"metrics-server-76d5d9b996-jzgqz\" (UID: \"169cb9e6-0d57-425d-a51d-a3b0e1556db6\") " pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.414213 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pf8kl\" (UniqueName: \"kubernetes.io/projected/169cb9e6-0d57-425d-a51d-a3b0e1556db6-kube-api-access-pf8kl\") pod \"metrics-server-76d5d9b996-jzgqz\" (UID: \"169cb9e6-0d57-425d-a51d-a3b0e1556db6\") " pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.414238 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/169cb9e6-0d57-425d-a51d-a3b0e1556db6-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-76d5d9b996-jzgqz\" (UID: \"169cb9e6-0d57-425d-a51d-a3b0e1556db6\") " pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.414268 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/169cb9e6-0d57-425d-a51d-a3b0e1556db6-metrics-server-audit-profiles\") pod \"metrics-server-76d5d9b996-jzgqz\" (UID: \"169cb9e6-0d57-425d-a51d-a3b0e1556db6\") " pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.414310 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/169cb9e6-0d57-425d-a51d-a3b0e1556db6-secret-metrics-server-tls\") pod \"metrics-server-76d5d9b996-jzgqz\" (UID: \"169cb9e6-0d57-425d-a51d-a3b0e1556db6\") " pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.414332 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/169cb9e6-0d57-425d-a51d-a3b0e1556db6-client-ca-bundle\") pod \"metrics-server-76d5d9b996-jzgqz\" (UID: \"169cb9e6-0d57-425d-a51d-a3b0e1556db6\") " pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.476147 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2f61301-d046-4651-b76c-9939b18fb792" path="/var/lib/kubelet/pods/c2f61301-d046-4651-b76c-9939b18fb792/volumes" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.514961 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/169cb9e6-0d57-425d-a51d-a3b0e1556db6-audit-log\") pod \"metrics-server-76d5d9b996-jzgqz\" (UID: \"169cb9e6-0d57-425d-a51d-a3b0e1556db6\") " pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.515024 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pf8kl\" (UniqueName: \"kubernetes.io/projected/169cb9e6-0d57-425d-a51d-a3b0e1556db6-kube-api-access-pf8kl\") pod \"metrics-server-76d5d9b996-jzgqz\" (UID: \"169cb9e6-0d57-425d-a51d-a3b0e1556db6\") " pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.515045 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/169cb9e6-0d57-425d-a51d-a3b0e1556db6-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-76d5d9b996-jzgqz\" (UID: \"169cb9e6-0d57-425d-a51d-a3b0e1556db6\") " pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.515070 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/169cb9e6-0d57-425d-a51d-a3b0e1556db6-metrics-server-audit-profiles\") pod \"metrics-server-76d5d9b996-jzgqz\" (UID: \"169cb9e6-0d57-425d-a51d-a3b0e1556db6\") " pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.515123 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/169cb9e6-0d57-425d-a51d-a3b0e1556db6-secret-metrics-server-tls\") pod \"metrics-server-76d5d9b996-jzgqz\" (UID: \"169cb9e6-0d57-425d-a51d-a3b0e1556db6\") " pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.515145 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/169cb9e6-0d57-425d-a51d-a3b0e1556db6-client-ca-bundle\") pod \"metrics-server-76d5d9b996-jzgqz\" (UID: \"169cb9e6-0d57-425d-a51d-a3b0e1556db6\") " pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.515179 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/169cb9e6-0d57-425d-a51d-a3b0e1556db6-secret-metrics-client-certs\") pod \"metrics-server-76d5d9b996-jzgqz\" (UID: \"169cb9e6-0d57-425d-a51d-a3b0e1556db6\") " pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.516552 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/169cb9e6-0d57-425d-a51d-a3b0e1556db6-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-76d5d9b996-jzgqz\" (UID: \"169cb9e6-0d57-425d-a51d-a3b0e1556db6\") " pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.516709 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/169cb9e6-0d57-425d-a51d-a3b0e1556db6-audit-log\") pod \"metrics-server-76d5d9b996-jzgqz\" (UID: \"169cb9e6-0d57-425d-a51d-a3b0e1556db6\") " pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.517616 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/169cb9e6-0d57-425d-a51d-a3b0e1556db6-metrics-server-audit-profiles\") pod \"metrics-server-76d5d9b996-jzgqz\" (UID: \"169cb9e6-0d57-425d-a51d-a3b0e1556db6\") " pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.524710 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/169cb9e6-0d57-425d-a51d-a3b0e1556db6-client-ca-bundle\") pod \"metrics-server-76d5d9b996-jzgqz\" (UID: \"169cb9e6-0d57-425d-a51d-a3b0e1556db6\") " pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.525394 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/169cb9e6-0d57-425d-a51d-a3b0e1556db6-secret-metrics-server-tls\") pod \"metrics-server-76d5d9b996-jzgqz\" (UID: \"169cb9e6-0d57-425d-a51d-a3b0e1556db6\") " pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.533865 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pf8kl\" (UniqueName: \"kubernetes.io/projected/169cb9e6-0d57-425d-a51d-a3b0e1556db6-kube-api-access-pf8kl\") pod \"metrics-server-76d5d9b996-jzgqz\" (UID: \"169cb9e6-0d57-425d-a51d-a3b0e1556db6\") " pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.534446 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/169cb9e6-0d57-425d-a51d-a3b0e1556db6-secret-metrics-client-certs\") pod \"metrics-server-76d5d9b996-jzgqz\" (UID: \"169cb9e6-0d57-425d-a51d-a3b0e1556db6\") " pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.739630 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.938468 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/monitoring-plugin-7d4f6975b-p9h8q"] Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.940580 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-7d4f6975b-p9h8q" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.944664 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"monitoring-plugin-cert" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.944805 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"default-dockercfg-6tstp" Dec 27 07:48:29 crc kubenswrapper[4934]: I1227 07:48:29.950832 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-7d4f6975b-p9h8q"] Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.021004 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/b78888a7-451e-43a3-86ea-b13c035afab1-monitoring-plugin-cert\") pod \"monitoring-plugin-7d4f6975b-p9h8q\" (UID: \"b78888a7-451e-43a3-86ea-b13c035afab1\") " pod="openshift-monitoring/monitoring-plugin-7d4f6975b-p9h8q" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.122787 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/b78888a7-451e-43a3-86ea-b13c035afab1-monitoring-plugin-cert\") pod \"monitoring-plugin-7d4f6975b-p9h8q\" (UID: \"b78888a7-451e-43a3-86ea-b13c035afab1\") " pod="openshift-monitoring/monitoring-plugin-7d4f6975b-p9h8q" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.138493 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/b78888a7-451e-43a3-86ea-b13c035afab1-monitoring-plugin-cert\") pod \"monitoring-plugin-7d4f6975b-p9h8q\" (UID: \"b78888a7-451e-43a3-86ea-b13c035afab1\") " pod="openshift-monitoring/monitoring-plugin-7d4f6975b-p9h8q" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.262293 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-7d4f6975b-p9h8q" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.269779 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb"] Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.269964 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" podUID="4598b565-6f32-486e-83ad-f72b288b5a38" containerName="route-controller-manager" containerID="cri-o://4750e5a042292fc070216f74a477b0be4a23812c65aca87e5ef8c6f02a567d35" gracePeriod=30 Dec 27 07:48:30 crc kubenswrapper[4934]: W1227 07:48:30.382141 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod772a3e80_411a_4c15_9e1c_b702392e13af.slice/crio-3e848cfc453369ea9b33d5ade5be9ab37a6f3b754a764cd1ae8606359b5d7297 WatchSource:0}: Error finding container 3e848cfc453369ea9b33d5ade5be9ab37a6f3b754a764cd1ae8606359b5d7297: Status 404 returned error can't find the container with id 3e848cfc453369ea9b33d5ade5be9ab37a6f3b754a764cd1ae8606359b5d7297 Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.631735 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.633683 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.636333 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"serving-certs-ca-bundle" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.636390 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.636501 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-kube-rbac-proxy-web" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.636564 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-web-config" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.636678 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-rbac-proxy" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.637696 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-prometheus-http-client-file" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.638446 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-sidecar-tls" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.638511 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-dockercfg-97dw5" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.639187 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.639724 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-k8s-rulefiles-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.639868 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls-assets-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.642242 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-grpc-tls-1140rbu3j95h5" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.645017 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-trusted-ca-bundle" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.660448 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.735813 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40c7cd23-14a7-4b30-9639-84d83ba74fe3-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.735879 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-web-config\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.735912 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.735943 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/40c7cd23-14a7-4b30-9639-84d83ba74fe3-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.735967 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/40c7cd23-14a7-4b30-9639-84d83ba74fe3-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.736159 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/40c7cd23-14a7-4b30-9639-84d83ba74fe3-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.736242 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.736288 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/40c7cd23-14a7-4b30-9639-84d83ba74fe3-config-out\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.736434 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.736458 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40c7cd23-14a7-4b30-9639-84d83ba74fe3-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.736484 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-config\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.736561 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ltdj\" (UniqueName: \"kubernetes.io/projected/40c7cd23-14a7-4b30-9639-84d83ba74fe3-kube-api-access-5ltdj\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.736623 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.736659 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.736748 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/40c7cd23-14a7-4b30-9639-84d83ba74fe3-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.736772 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.736816 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.736860 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40c7cd23-14a7-4b30-9639-84d83ba74fe3-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.838276 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-config\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.838322 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ltdj\" (UniqueName: \"kubernetes.io/projected/40c7cd23-14a7-4b30-9639-84d83ba74fe3-kube-api-access-5ltdj\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.838359 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.838378 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.838404 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.838420 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/40c7cd23-14a7-4b30-9639-84d83ba74fe3-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.838437 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.838460 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40c7cd23-14a7-4b30-9639-84d83ba74fe3-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.838487 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40c7cd23-14a7-4b30-9639-84d83ba74fe3-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.838508 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-web-config\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.838527 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.838545 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/40c7cd23-14a7-4b30-9639-84d83ba74fe3-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.838568 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/40c7cd23-14a7-4b30-9639-84d83ba74fe3-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.838591 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/40c7cd23-14a7-4b30-9639-84d83ba74fe3-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.838614 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.838631 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/40c7cd23-14a7-4b30-9639-84d83ba74fe3-config-out\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.838653 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.838668 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40c7cd23-14a7-4b30-9639-84d83ba74fe3-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.839457 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40c7cd23-14a7-4b30-9639-84d83ba74fe3-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.839764 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/40c7cd23-14a7-4b30-9639-84d83ba74fe3-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.840341 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/40c7cd23-14a7-4b30-9639-84d83ba74fe3-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.840914 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40c7cd23-14a7-4b30-9639-84d83ba74fe3-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.843248 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-config\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.843444 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.843945 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/40c7cd23-14a7-4b30-9639-84d83ba74fe3-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.844456 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-web-config\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.844865 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/40c7cd23-14a7-4b30-9639-84d83ba74fe3-config-out\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.845862 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40c7cd23-14a7-4b30-9639-84d83ba74fe3-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.845868 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.846241 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.849609 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.849664 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.852393 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.859300 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/40c7cd23-14a7-4b30-9639-84d83ba74fe3-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.867830 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ltdj\" (UniqueName: \"kubernetes.io/projected/40c7cd23-14a7-4b30-9639-84d83ba74fe3-kube-api-access-5ltdj\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.869141 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/40c7cd23-14a7-4b30-9639-84d83ba74fe3-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"40c7cd23-14a7-4b30-9639-84d83ba74fe3\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:30 crc kubenswrapper[4934]: I1227 07:48:30.951853 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.188692 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" event={"ID":"772a3e80-411a-4c15-9e1c-b702392e13af","Type":"ContainerStarted","Data":"3e848cfc453369ea9b33d5ade5be9ab37a6f3b754a764cd1ae8606359b5d7297"} Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.190596 4934 generic.go:334] "Generic (PLEG): container finished" podID="4598b565-6f32-486e-83ad-f72b288b5a38" containerID="4750e5a042292fc070216f74a477b0be4a23812c65aca87e5ef8c6f02a567d35" exitCode=0 Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.190657 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" event={"ID":"4598b565-6f32-486e-83ad-f72b288b5a38","Type":"ContainerDied","Data":"4750e5a042292fc070216f74a477b0be4a23812c65aca87e5ef8c6f02a567d35"} Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.643666 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.667895 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz"] Dec 27 07:48:31 crc kubenswrapper[4934]: E1227 07:48:31.668146 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4598b565-6f32-486e-83ad-f72b288b5a38" containerName="route-controller-manager" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.668161 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="4598b565-6f32-486e-83ad-f72b288b5a38" containerName="route-controller-manager" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.668253 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="4598b565-6f32-486e-83ad-f72b288b5a38" containerName="route-controller-manager" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.668642 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.674867 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz"] Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.761442 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4598b565-6f32-486e-83ad-f72b288b5a38-serving-cert\") pod \"4598b565-6f32-486e-83ad-f72b288b5a38\" (UID: \"4598b565-6f32-486e-83ad-f72b288b5a38\") " Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.761549 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4598b565-6f32-486e-83ad-f72b288b5a38-config\") pod \"4598b565-6f32-486e-83ad-f72b288b5a38\" (UID: \"4598b565-6f32-486e-83ad-f72b288b5a38\") " Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.761583 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9mtm\" (UniqueName: \"kubernetes.io/projected/4598b565-6f32-486e-83ad-f72b288b5a38-kube-api-access-c9mtm\") pod \"4598b565-6f32-486e-83ad-f72b288b5a38\" (UID: \"4598b565-6f32-486e-83ad-f72b288b5a38\") " Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.761625 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4598b565-6f32-486e-83ad-f72b288b5a38-client-ca\") pod \"4598b565-6f32-486e-83ad-f72b288b5a38\" (UID: \"4598b565-6f32-486e-83ad-f72b288b5a38\") " Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.762296 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4598b565-6f32-486e-83ad-f72b288b5a38-client-ca" (OuterVolumeSpecName: "client-ca") pod "4598b565-6f32-486e-83ad-f72b288b5a38" (UID: "4598b565-6f32-486e-83ad-f72b288b5a38"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.762345 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4598b565-6f32-486e-83ad-f72b288b5a38-config" (OuterVolumeSpecName: "config") pod "4598b565-6f32-486e-83ad-f72b288b5a38" (UID: "4598b565-6f32-486e-83ad-f72b288b5a38"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.762561 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45f4ca46-4041-4960-8c05-a20bd2c90b90-config\") pod \"route-controller-manager-6bd9bdbfbf-x5vvz\" (UID: \"45f4ca46-4041-4960-8c05-a20bd2c90b90\") " pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.762632 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45f4ca46-4041-4960-8c05-a20bd2c90b90-serving-cert\") pod \"route-controller-manager-6bd9bdbfbf-x5vvz\" (UID: \"45f4ca46-4041-4960-8c05-a20bd2c90b90\") " pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.762653 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45f4ca46-4041-4960-8c05-a20bd2c90b90-client-ca\") pod \"route-controller-manager-6bd9bdbfbf-x5vvz\" (UID: \"45f4ca46-4041-4960-8c05-a20bd2c90b90\") " pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.762746 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q68g4\" (UniqueName: \"kubernetes.io/projected/45f4ca46-4041-4960-8c05-a20bd2c90b90-kube-api-access-q68g4\") pod \"route-controller-manager-6bd9bdbfbf-x5vvz\" (UID: \"45f4ca46-4041-4960-8c05-a20bd2c90b90\") " pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.762815 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4598b565-6f32-486e-83ad-f72b288b5a38-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.762826 4934 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4598b565-6f32-486e-83ad-f72b288b5a38-client-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.772679 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4598b565-6f32-486e-83ad-f72b288b5a38-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4598b565-6f32-486e-83ad-f72b288b5a38" (UID: "4598b565-6f32-486e-83ad-f72b288b5a38"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.773296 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4598b565-6f32-486e-83ad-f72b288b5a38-kube-api-access-c9mtm" (OuterVolumeSpecName: "kube-api-access-c9mtm") pod "4598b565-6f32-486e-83ad-f72b288b5a38" (UID: "4598b565-6f32-486e-83ad-f72b288b5a38"). InnerVolumeSpecName "kube-api-access-c9mtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.796388 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-76d5d9b996-jzgqz"] Dec 27 07:48:31 crc kubenswrapper[4934]: W1227 07:48:31.803023 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod169cb9e6_0d57_425d_a51d_a3b0e1556db6.slice/crio-e504d543431e6727c00af4e33495c824f53d8b74046dc9e38792cf3388236ac7 WatchSource:0}: Error finding container e504d543431e6727c00af4e33495c824f53d8b74046dc9e38792cf3388236ac7: Status 404 returned error can't find the container with id e504d543431e6727c00af4e33495c824f53d8b74046dc9e38792cf3388236ac7 Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.859228 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-c8c89d49c-b97s8"] Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.864134 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q68g4\" (UniqueName: \"kubernetes.io/projected/45f4ca46-4041-4960-8c05-a20bd2c90b90-kube-api-access-q68g4\") pod \"route-controller-manager-6bd9bdbfbf-x5vvz\" (UID: \"45f4ca46-4041-4960-8c05-a20bd2c90b90\") " pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.864183 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45f4ca46-4041-4960-8c05-a20bd2c90b90-config\") pod \"route-controller-manager-6bd9bdbfbf-x5vvz\" (UID: \"45f4ca46-4041-4960-8c05-a20bd2c90b90\") " pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.864216 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45f4ca46-4041-4960-8c05-a20bd2c90b90-serving-cert\") pod \"route-controller-manager-6bd9bdbfbf-x5vvz\" (UID: \"45f4ca46-4041-4960-8c05-a20bd2c90b90\") " pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.864243 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45f4ca46-4041-4960-8c05-a20bd2c90b90-client-ca\") pod \"route-controller-manager-6bd9bdbfbf-x5vvz\" (UID: \"45f4ca46-4041-4960-8c05-a20bd2c90b90\") " pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.864287 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4598b565-6f32-486e-83ad-f72b288b5a38-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.864300 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9mtm\" (UniqueName: \"kubernetes.io/projected/4598b565-6f32-486e-83ad-f72b288b5a38-kube-api-access-c9mtm\") on node \"crc\" DevicePath \"\"" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.865109 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45f4ca46-4041-4960-8c05-a20bd2c90b90-client-ca\") pod \"route-controller-manager-6bd9bdbfbf-x5vvz\" (UID: \"45f4ca46-4041-4960-8c05-a20bd2c90b90\") " pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.866699 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45f4ca46-4041-4960-8c05-a20bd2c90b90-config\") pod \"route-controller-manager-6bd9bdbfbf-x5vvz\" (UID: \"45f4ca46-4041-4960-8c05-a20bd2c90b90\") " pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.869165 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45f4ca46-4041-4960-8c05-a20bd2c90b90-serving-cert\") pod \"route-controller-manager-6bd9bdbfbf-x5vvz\" (UID: \"45f4ca46-4041-4960-8c05-a20bd2c90b90\") " pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" Dec 27 07:48:31 crc kubenswrapper[4934]: W1227 07:48:31.877365 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6aa1cec9_a0f2_46da_ae87_b94936a1360b.slice/crio-a72702fbb69e3d117de98b84849a14e24c453d2440e5cdaf3901d3905dd77c92 WatchSource:0}: Error finding container a72702fbb69e3d117de98b84849a14e24c453d2440e5cdaf3901d3905dd77c92: Status 404 returned error can't find the container with id a72702fbb69e3d117de98b84849a14e24c453d2440e5cdaf3901d3905dd77c92 Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.877647 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.891941 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q68g4\" (UniqueName: \"kubernetes.io/projected/45f4ca46-4041-4960-8c05-a20bd2c90b90-kube-api-access-q68g4\") pod \"route-controller-manager-6bd9bdbfbf-x5vvz\" (UID: \"45f4ca46-4041-4960-8c05-a20bd2c90b90\") " pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.985779 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-7d4f6975b-p9h8q"] Dec 27 07:48:31 crc kubenswrapper[4934]: I1227 07:48:31.990492 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.200974 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" event={"ID":"169cb9e6-0d57-425d-a51d-a3b0e1556db6","Type":"ContainerStarted","Data":"e504d543431e6727c00af4e33495c824f53d8b74046dc9e38792cf3388236ac7"} Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.206029 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"17fe170d-1e03-46fb-a0ea-3b209cb114ff","Type":"ContainerStarted","Data":"f8d3146b14571bd972dbd447996a99cdb8d81b5efea006c6ef4f48a8bc2c88ca"} Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.206086 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"17fe170d-1e03-46fb-a0ea-3b209cb114ff","Type":"ContainerStarted","Data":"38ff8ed1f4c50959be2d39ba78d40bb6e4c63201c8cecacff5e84dea469ded3b"} Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.206119 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"17fe170d-1e03-46fb-a0ea-3b209cb114ff","Type":"ContainerStarted","Data":"82b21dbc142ed416174f8d0445c3c79e02dc21258994b0ac15c749a9d27d1d29"} Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.206129 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"17fe170d-1e03-46fb-a0ea-3b209cb114ff","Type":"ContainerStarted","Data":"fbef17ff8b8d20866d6652a714fae2e655267ab734d18c446a3f5a1e204c9788"} Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.207880 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" event={"ID":"4598b565-6f32-486e-83ad-f72b288b5a38","Type":"ContainerDied","Data":"af428d12d26c7d9e74deae0f6995710ca1ee9bab4c22022b8abef66bb340aa4a"} Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.207950 4934 scope.go:117] "RemoveContainer" containerID="4750e5a042292fc070216f74a477b0be4a23812c65aca87e5ef8c6f02a567d35" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.207986 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.211043 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-8467887986-svb7h" event={"ID":"a07fdfc4-584a-4a39-b87f-ee78f9cac687","Type":"ContainerStarted","Data":"7127056b608dfd4aec0a9c8f3fb868cb5b1905ba6470ab8596b6e2d7550b7316"} Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.211096 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-8467887986-svb7h" event={"ID":"a07fdfc4-584a-4a39-b87f-ee78f9cac687","Type":"ContainerStarted","Data":"6b752341e86cd3b99cab99398f9bd8cf1dae19c603d03b7027b53c0225c826d5"} Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.211107 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-8467887986-svb7h" event={"ID":"a07fdfc4-584a-4a39-b87f-ee78f9cac687","Type":"ContainerStarted","Data":"308b735acfa8776de1ff4c4e94c5ba642714d1611ce16883016ad3cc9b2381c9"} Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.212508 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-c8c89d49c-b97s8" event={"ID":"6aa1cec9-a0f2-46da-ae87-b94936a1360b","Type":"ContainerStarted","Data":"b581d03cb9679949a3f4c7b52b451ceda693e476513699728786d21b2e99a932"} Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.212537 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-c8c89d49c-b97s8" event={"ID":"6aa1cec9-a0f2-46da-ae87-b94936a1360b","Type":"ContainerStarted","Data":"a72702fbb69e3d117de98b84849a14e24c453d2440e5cdaf3901d3905dd77c92"} Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.215386 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-7d4f6975b-p9h8q" event={"ID":"b78888a7-451e-43a3-86ea-b13c035afab1","Type":"ContainerStarted","Data":"fd2005ea03fe353c15588a8a1d480db39b5147e25787a530ac399467c7f80ad3"} Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.216844 4934 generic.go:334] "Generic (PLEG): container finished" podID="40c7cd23-14a7-4b30-9639-84d83ba74fe3" containerID="a3805b630ca7d9557c2bbcaa40d72fa6bfaa4f2a9000a1293d2f220230ab990a" exitCode=0 Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.216895 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"40c7cd23-14a7-4b30-9639-84d83ba74fe3","Type":"ContainerDied","Data":"a3805b630ca7d9557c2bbcaa40d72fa6bfaa4f2a9000a1293d2f220230ab990a"} Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.216918 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"40c7cd23-14a7-4b30-9639-84d83ba74fe3","Type":"ContainerStarted","Data":"89a68a40d1a37e77b0a0f105a0096527151957f0fed77e9182eaa5d1741f153d"} Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.218771 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" event={"ID":"772a3e80-411a-4c15-9e1c-b702392e13af","Type":"ContainerStarted","Data":"f9908286ac4156a77e85f0cdcafc047330f167accce9c72fcb833f0c42c2b04e"} Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.219434 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.237417 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-c8c89d49c-b97s8" podStartSLOduration=4.237401402 podStartE2EDuration="4.237401402s" podCreationTimestamp="2025-12-27 07:48:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:48:32.233630439 +0000 UTC m=+373.054071043" watchObservedRunningTime="2025-12-27 07:48:32.237401402 +0000 UTC m=+373.057841996" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.260757 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" podStartSLOduration=30.260734557 podStartE2EDuration="30.260734557s" podCreationTimestamp="2025-12-27 07:48:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:48:32.254927729 +0000 UTC m=+373.075368333" watchObservedRunningTime="2025-12-27 07:48:32.260734557 +0000 UTC m=+373.081175151" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.267276 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb"] Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.270502 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.273044 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5fdfb79994-qlktb"] Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.421536 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz"] Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.844489 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-vkfr4"] Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.845613 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.873893 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-vkfr4"] Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.885777 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/21dab295-f1e7-4fee-a281-251122757a57-registry-tls\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.885891 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gttr\" (UniqueName: \"kubernetes.io/projected/21dab295-f1e7-4fee-a281-251122757a57-kube-api-access-2gttr\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.886108 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/21dab295-f1e7-4fee-a281-251122757a57-installation-pull-secrets\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.886156 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/21dab295-f1e7-4fee-a281-251122757a57-registry-certificates\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.886297 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/21dab295-f1e7-4fee-a281-251122757a57-bound-sa-token\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.886343 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/21dab295-f1e7-4fee-a281-251122757a57-trusted-ca\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.886387 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/21dab295-f1e7-4fee-a281-251122757a57-ca-trust-extracted\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.886417 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.916543 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.988172 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/21dab295-f1e7-4fee-a281-251122757a57-bound-sa-token\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.988239 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/21dab295-f1e7-4fee-a281-251122757a57-trusted-ca\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.988282 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/21dab295-f1e7-4fee-a281-251122757a57-ca-trust-extracted\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.988340 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/21dab295-f1e7-4fee-a281-251122757a57-registry-tls\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.988375 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gttr\" (UniqueName: \"kubernetes.io/projected/21dab295-f1e7-4fee-a281-251122757a57-kube-api-access-2gttr\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.988421 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/21dab295-f1e7-4fee-a281-251122757a57-installation-pull-secrets\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.988448 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/21dab295-f1e7-4fee-a281-251122757a57-registry-certificates\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.989601 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/21dab295-f1e7-4fee-a281-251122757a57-trusted-ca\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.989952 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/21dab295-f1e7-4fee-a281-251122757a57-ca-trust-extracted\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.990286 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/21dab295-f1e7-4fee-a281-251122757a57-registry-certificates\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.995207 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/21dab295-f1e7-4fee-a281-251122757a57-installation-pull-secrets\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:32 crc kubenswrapper[4934]: I1227 07:48:32.995211 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/21dab295-f1e7-4fee-a281-251122757a57-registry-tls\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:33 crc kubenswrapper[4934]: I1227 07:48:33.004425 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/21dab295-f1e7-4fee-a281-251122757a57-bound-sa-token\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:33 crc kubenswrapper[4934]: I1227 07:48:33.008929 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gttr\" (UniqueName: \"kubernetes.io/projected/21dab295-f1e7-4fee-a281-251122757a57-kube-api-access-2gttr\") pod \"image-registry-66df7c8f76-vkfr4\" (UID: \"21dab295-f1e7-4fee-a281-251122757a57\") " pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:33 crc kubenswrapper[4934]: I1227 07:48:33.175063 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:33 crc kubenswrapper[4934]: I1227 07:48:33.226798 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" event={"ID":"45f4ca46-4041-4960-8c05-a20bd2c90b90","Type":"ContainerStarted","Data":"5f66e5c0191c6ce691d9d5526a2dabacd95623f4549ca175a37bdd114a0a0a6d"} Dec 27 07:48:33 crc kubenswrapper[4934]: I1227 07:48:33.226861 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" event={"ID":"45f4ca46-4041-4960-8c05-a20bd2c90b90","Type":"ContainerStarted","Data":"a74f53862662140eb2148708fd857234f71a4da5e462065fe2df7db4fab7ed34"} Dec 27 07:48:33 crc kubenswrapper[4934]: I1227 07:48:33.229516 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" Dec 27 07:48:33 crc kubenswrapper[4934]: I1227 07:48:33.234783 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" Dec 27 07:48:33 crc kubenswrapper[4934]: I1227 07:48:33.235939 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"17fe170d-1e03-46fb-a0ea-3b209cb114ff","Type":"ContainerStarted","Data":"fee4b5082abc4c33d76e5d1b127feb8c2b8c8a9c2d72c5c29d8ded17d78b5a84"} Dec 27 07:48:33 crc kubenswrapper[4934]: I1227 07:48:33.253116 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" podStartSLOduration=3.253060534 podStartE2EDuration="3.253060534s" podCreationTimestamp="2025-12-27 07:48:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:48:33.250505877 +0000 UTC m=+374.070946471" watchObservedRunningTime="2025-12-27 07:48:33.253060534 +0000 UTC m=+374.073501128" Dec 27 07:48:33 crc kubenswrapper[4934]: I1227 07:48:33.480028 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4598b565-6f32-486e-83ad-f72b288b5a38" path="/var/lib/kubelet/pods/4598b565-6f32-486e-83ad-f72b288b5a38/volumes" Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.055521 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qtq22"] Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.057180 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qtq22" Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.060160 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.060547 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qtq22"] Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.148558 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-vkfr4"] Dec 27 07:48:36 crc kubenswrapper[4934]: W1227 07:48:36.150648 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21dab295_f1e7_4fee_a281_251122757a57.slice/crio-aafbf562c737f8e1367b40dfcaea2b8a6ee4c765092e58f2a1cf9ab9e28193ee WatchSource:0}: Error finding container aafbf562c737f8e1367b40dfcaea2b8a6ee4c765092e58f2a1cf9ab9e28193ee: Status 404 returned error can't find the container with id aafbf562c737f8e1367b40dfcaea2b8a6ee4c765092e58f2a1cf9ab9e28193ee Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.153047 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ee60051-8eec-4da3-ae78-b409c5cdf9f3-utilities\") pod \"redhat-operators-qtq22\" (UID: \"2ee60051-8eec-4da3-ae78-b409c5cdf9f3\") " pod="openshift-marketplace/redhat-operators-qtq22" Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.153409 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ee60051-8eec-4da3-ae78-b409c5cdf9f3-catalog-content\") pod \"redhat-operators-qtq22\" (UID: \"2ee60051-8eec-4da3-ae78-b409c5cdf9f3\") " pod="openshift-marketplace/redhat-operators-qtq22" Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.153436 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbsl5\" (UniqueName: \"kubernetes.io/projected/2ee60051-8eec-4da3-ae78-b409c5cdf9f3-kube-api-access-pbsl5\") pod \"redhat-operators-qtq22\" (UID: \"2ee60051-8eec-4da3-ae78-b409c5cdf9f3\") " pod="openshift-marketplace/redhat-operators-qtq22" Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.254399 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ee60051-8eec-4da3-ae78-b409c5cdf9f3-catalog-content\") pod \"redhat-operators-qtq22\" (UID: \"2ee60051-8eec-4da3-ae78-b409c5cdf9f3\") " pod="openshift-marketplace/redhat-operators-qtq22" Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.254728 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbsl5\" (UniqueName: \"kubernetes.io/projected/2ee60051-8eec-4da3-ae78-b409c5cdf9f3-kube-api-access-pbsl5\") pod \"redhat-operators-qtq22\" (UID: \"2ee60051-8eec-4da3-ae78-b409c5cdf9f3\") " pod="openshift-marketplace/redhat-operators-qtq22" Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.254804 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ee60051-8eec-4da3-ae78-b409c5cdf9f3-utilities\") pod \"redhat-operators-qtq22\" (UID: \"2ee60051-8eec-4da3-ae78-b409c5cdf9f3\") " pod="openshift-marketplace/redhat-operators-qtq22" Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.254945 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ee60051-8eec-4da3-ae78-b409c5cdf9f3-catalog-content\") pod \"redhat-operators-qtq22\" (UID: \"2ee60051-8eec-4da3-ae78-b409c5cdf9f3\") " pod="openshift-marketplace/redhat-operators-qtq22" Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.255181 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ee60051-8eec-4da3-ae78-b409c5cdf9f3-utilities\") pod \"redhat-operators-qtq22\" (UID: \"2ee60051-8eec-4da3-ae78-b409c5cdf9f3\") " pod="openshift-marketplace/redhat-operators-qtq22" Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.256709 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"17fe170d-1e03-46fb-a0ea-3b209cb114ff","Type":"ContainerStarted","Data":"7852ed8257fc157bc1a6ca5906681a30655b83d51a02cba216890636a3d749e3"} Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.265372 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-8467887986-svb7h" event={"ID":"a07fdfc4-584a-4a39-b87f-ee78f9cac687","Type":"ContainerStarted","Data":"75c8046b281910d4648902701d61093a5f27920077e7c84c4833f2ea4bbd15d1"} Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.265436 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-8467887986-svb7h" event={"ID":"a07fdfc4-584a-4a39-b87f-ee78f9cac687","Type":"ContainerStarted","Data":"311a8d2a0a0c1fac1e419f656bfa49622c0d4f17b3c21c0a2b1d462263ed0cdf"} Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.266922 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-7d4f6975b-p9h8q" event={"ID":"b78888a7-451e-43a3-86ea-b13c035afab1","Type":"ContainerStarted","Data":"9203f9fab60001ce255a872447663970d749255507316c53c40dfbc4946f5c23"} Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.267291 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/monitoring-plugin-7d4f6975b-p9h8q" Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.268735 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"40c7cd23-14a7-4b30-9639-84d83ba74fe3","Type":"ContainerStarted","Data":"2a6a7441808fcfcb26cc84fb326cb0fe7415db8d2618a851cdfbd53934884b7d"} Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.268770 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"40c7cd23-14a7-4b30-9639-84d83ba74fe3","Type":"ContainerStarted","Data":"8e14c0b38942133295106d0a988bdbf29a220d18b45bf40a743baf8bc0700770"} Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.269949 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" event={"ID":"169cb9e6-0d57-425d-a51d-a3b0e1556db6","Type":"ContainerStarted","Data":"e40036310a422895aa9efdc01f5650f96be142fb88c8c8114bfb66e87396cf64"} Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.271416 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" event={"ID":"21dab295-f1e7-4fee-a281-251122757a57","Type":"ContainerStarted","Data":"aafbf562c737f8e1367b40dfcaea2b8a6ee4c765092e58f2a1cf9ab9e28193ee"} Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.276873 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbsl5\" (UniqueName: \"kubernetes.io/projected/2ee60051-8eec-4da3-ae78-b409c5cdf9f3-kube-api-access-pbsl5\") pod \"redhat-operators-qtq22\" (UID: \"2ee60051-8eec-4da3-ae78-b409c5cdf9f3\") " pod="openshift-marketplace/redhat-operators-qtq22" Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.281252 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/monitoring-plugin-7d4f6975b-p9h8q" Dec 27 07:48:36 crc kubenswrapper[4934]: I1227 07:48:36.290353 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=1.9874290079999999 podStartE2EDuration="11.290335596s" podCreationTimestamp="2025-12-27 07:48:25 +0000 UTC" firstStartedPulling="2025-12-27 07:48:26.38183773 +0000 UTC m=+367.202278364" lastFinishedPulling="2025-12-27 07:48:35.684744358 +0000 UTC m=+376.505184952" observedRunningTime="2025-12-27 07:48:36.285408557 +0000 UTC m=+377.105849171" watchObservedRunningTime="2025-12-27 07:48:36.290335596 +0000 UTC m=+377.110776190" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:36.342788 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" podStartSLOduration=3.402452795 podStartE2EDuration="7.342768154s" podCreationTimestamp="2025-12-27 07:48:29 +0000 UTC" firstStartedPulling="2025-12-27 07:48:31.809039455 +0000 UTC m=+372.629480049" lastFinishedPulling="2025-12-27 07:48:35.749354814 +0000 UTC m=+376.569795408" observedRunningTime="2025-12-27 07:48:36.320654656 +0000 UTC m=+377.141095270" watchObservedRunningTime="2025-12-27 07:48:36.342768154 +0000 UTC m=+377.163208748" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:36.343156 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/monitoring-plugin-7d4f6975b-p9h8q" podStartSLOduration=3.5897030279999997 podStartE2EDuration="7.343150882s" podCreationTimestamp="2025-12-27 07:48:29 +0000 UTC" firstStartedPulling="2025-12-27 07:48:32.00631739 +0000 UTC m=+372.826757984" lastFinishedPulling="2025-12-27 07:48:35.759765244 +0000 UTC m=+376.580205838" observedRunningTime="2025-12-27 07:48:36.34258664 +0000 UTC m=+377.163027244" watchObservedRunningTime="2025-12-27 07:48:36.343150882 +0000 UTC m=+377.163591486" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:36.389292 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qtq22" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:36.660761 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lbfd2"] Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:36.662447 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lbfd2" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:36.663854 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:36.671642 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lbfd2"] Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:36.765740 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0661ef0e-af9d-41ba-8059-2e878b11b042-utilities\") pod \"certified-operators-lbfd2\" (UID: \"0661ef0e-af9d-41ba-8059-2e878b11b042\") " pod="openshift-marketplace/certified-operators-lbfd2" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:36.765792 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0661ef0e-af9d-41ba-8059-2e878b11b042-catalog-content\") pod \"certified-operators-lbfd2\" (UID: \"0661ef0e-af9d-41ba-8059-2e878b11b042\") " pod="openshift-marketplace/certified-operators-lbfd2" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:36.765997 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwvj7\" (UniqueName: \"kubernetes.io/projected/0661ef0e-af9d-41ba-8059-2e878b11b042-kube-api-access-zwvj7\") pod \"certified-operators-lbfd2\" (UID: \"0661ef0e-af9d-41ba-8059-2e878b11b042\") " pod="openshift-marketplace/certified-operators-lbfd2" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:36.867812 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0661ef0e-af9d-41ba-8059-2e878b11b042-catalog-content\") pod \"certified-operators-lbfd2\" (UID: \"0661ef0e-af9d-41ba-8059-2e878b11b042\") " pod="openshift-marketplace/certified-operators-lbfd2" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:36.868262 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwvj7\" (UniqueName: \"kubernetes.io/projected/0661ef0e-af9d-41ba-8059-2e878b11b042-kube-api-access-zwvj7\") pod \"certified-operators-lbfd2\" (UID: \"0661ef0e-af9d-41ba-8059-2e878b11b042\") " pod="openshift-marketplace/certified-operators-lbfd2" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:36.868364 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0661ef0e-af9d-41ba-8059-2e878b11b042-utilities\") pod \"certified-operators-lbfd2\" (UID: \"0661ef0e-af9d-41ba-8059-2e878b11b042\") " pod="openshift-marketplace/certified-operators-lbfd2" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:36.868471 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0661ef0e-af9d-41ba-8059-2e878b11b042-catalog-content\") pod \"certified-operators-lbfd2\" (UID: \"0661ef0e-af9d-41ba-8059-2e878b11b042\") " pod="openshift-marketplace/certified-operators-lbfd2" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:36.868714 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0661ef0e-af9d-41ba-8059-2e878b11b042-utilities\") pod \"certified-operators-lbfd2\" (UID: \"0661ef0e-af9d-41ba-8059-2e878b11b042\") " pod="openshift-marketplace/certified-operators-lbfd2" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:36.890049 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwvj7\" (UniqueName: \"kubernetes.io/projected/0661ef0e-af9d-41ba-8059-2e878b11b042-kube-api-access-zwvj7\") pod \"certified-operators-lbfd2\" (UID: \"0661ef0e-af9d-41ba-8059-2e878b11b042\") " pod="openshift-marketplace/certified-operators-lbfd2" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:37.006693 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lbfd2" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:37.276925 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" event={"ID":"21dab295-f1e7-4fee-a281-251122757a57","Type":"ContainerStarted","Data":"f539ef01d230f03785b6cd41476e80595a613fd5ebc5f68de4b256b9267f5fb4"} Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:37.277257 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:37.288108 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-8467887986-svb7h" event={"ID":"a07fdfc4-584a-4a39-b87f-ee78f9cac687","Type":"ContainerStarted","Data":"5f76e88a4c55f8cdfa0070084b794327aed8bc61bf6cab0ba6b886b61dc8ae7d"} Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:37.288938 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:37.291400 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_40c7cd23-14a7-4b30-9639-84d83ba74fe3/prometheus/0.log" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:37.291792 4934 generic.go:334] "Generic (PLEG): container finished" podID="40c7cd23-14a7-4b30-9639-84d83ba74fe3" containerID="8e14c0b38942133295106d0a988bdbf29a220d18b45bf40a743baf8bc0700770" exitCode=1 Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:37.292179 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"40c7cd23-14a7-4b30-9639-84d83ba74fe3","Type":"ContainerDied","Data":"8e14c0b38942133295106d0a988bdbf29a220d18b45bf40a743baf8bc0700770"} Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:37.292200 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"40c7cd23-14a7-4b30-9639-84d83ba74fe3","Type":"ContainerStarted","Data":"5844e99a2aa4b9644d1c95ad3fb217cf4fc7a6c2742307344229666e4626ee2d"} Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:37.292210 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"40c7cd23-14a7-4b30-9639-84d83ba74fe3","Type":"ContainerStarted","Data":"21de0e586ffcc44cc12fbd2c651db50ce1f7369c8cc9f608ef96ec93cef092d1"} Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:37.292219 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"40c7cd23-14a7-4b30-9639-84d83ba74fe3","Type":"ContainerStarted","Data":"7c9d3978da8a7934457e9d74a83b34a5299f0dff1aea81e66ba8fc9d59c560f2"} Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:37.292228 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"40c7cd23-14a7-4b30-9639-84d83ba74fe3","Type":"ContainerStarted","Data":"c2ca9a0ff91b526dbc136da983c8b40b3010c9d17f68bd6511127a56dc41fe2e"} Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:37.292947 4934 scope.go:117] "RemoveContainer" containerID="8e14c0b38942133295106d0a988bdbf29a220d18b45bf40a743baf8bc0700770" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:37.302507 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-8467887986-svb7h" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:37.334161 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" podStartSLOduration=5.33413801 podStartE2EDuration="5.33413801s" podCreationTimestamp="2025-12-27 07:48:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:48:37.296870347 +0000 UTC m=+378.117310941" watchObservedRunningTime="2025-12-27 07:48:37.33413801 +0000 UTC m=+378.154578604" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:37.360154 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/thanos-querier-8467887986-svb7h" podStartSLOduration=2.600806498 podStartE2EDuration="11.360131344s" podCreationTimestamp="2025-12-27 07:48:26 +0000 UTC" firstStartedPulling="2025-12-27 07:48:26.925409531 +0000 UTC m=+367.745850125" lastFinishedPulling="2025-12-27 07:48:35.684734377 +0000 UTC m=+376.505174971" observedRunningTime="2025-12-27 07:48:37.349621442 +0000 UTC m=+378.170062056" watchObservedRunningTime="2025-12-27 07:48:37.360131344 +0000 UTC m=+378.180571948" Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:37.767741 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lbfd2"] Dec 27 07:48:37 crc kubenswrapper[4934]: W1227 07:48:37.774810 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0661ef0e_af9d_41ba_8059_2e878b11b042.slice/crio-4791843c23c9d6750e6cb0f977f6b7833929766197580316c042766b872e3509 WatchSource:0}: Error finding container 4791843c23c9d6750e6cb0f977f6b7833929766197580316c042766b872e3509: Status 404 returned error can't find the container with id 4791843c23c9d6750e6cb0f977f6b7833929766197580316c042766b872e3509 Dec 27 07:48:37 crc kubenswrapper[4934]: I1227 07:48:37.797511 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qtq22"] Dec 27 07:48:37 crc kubenswrapper[4934]: W1227 07:48:37.801594 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ee60051_8eec_4da3_ae78_b409c5cdf9f3.slice/crio-1eaa4097b44192e81d6544f28960aaed17e2a122a5a26973ed8557daf0c2b248 WatchSource:0}: Error finding container 1eaa4097b44192e81d6544f28960aaed17e2a122a5a26973ed8557daf0c2b248: Status 404 returned error can't find the container with id 1eaa4097b44192e81d6544f28960aaed17e2a122a5a26973ed8557daf0c2b248 Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.302164 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_40c7cd23-14a7-4b30-9639-84d83ba74fe3/prometheus/1.log" Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.305074 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_40c7cd23-14a7-4b30-9639-84d83ba74fe3/prometheus/0.log" Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.305537 4934 generic.go:334] "Generic (PLEG): container finished" podID="40c7cd23-14a7-4b30-9639-84d83ba74fe3" containerID="41c879bc517aa226186963efe3788cca745786da2466030eb60387789ec75d06" exitCode=1 Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.305611 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"40c7cd23-14a7-4b30-9639-84d83ba74fe3","Type":"ContainerDied","Data":"41c879bc517aa226186963efe3788cca745786da2466030eb60387789ec75d06"} Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.305649 4934 scope.go:117] "RemoveContainer" containerID="8e14c0b38942133295106d0a988bdbf29a220d18b45bf40a743baf8bc0700770" Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.306940 4934 scope.go:117] "RemoveContainer" containerID="41c879bc517aa226186963efe3788cca745786da2466030eb60387789ec75d06" Dec 27 07:48:38 crc kubenswrapper[4934]: E1227 07:48:38.307435 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=prometheus pod=prometheus-k8s-0_openshift-monitoring(40c7cd23-14a7-4b30-9639-84d83ba74fe3)\"" pod="openshift-monitoring/prometheus-k8s-0" podUID="40c7cd23-14a7-4b30-9639-84d83ba74fe3" Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.308030 4934 generic.go:334] "Generic (PLEG): container finished" podID="0661ef0e-af9d-41ba-8059-2e878b11b042" containerID="34dee72fc416509b19ea359cefab26c36abc7056cb06baa9a2c4711f1f2d6fbe" exitCode=0 Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.308139 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lbfd2" event={"ID":"0661ef0e-af9d-41ba-8059-2e878b11b042","Type":"ContainerDied","Data":"34dee72fc416509b19ea359cefab26c36abc7056cb06baa9a2c4711f1f2d6fbe"} Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.308176 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lbfd2" event={"ID":"0661ef0e-af9d-41ba-8059-2e878b11b042","Type":"ContainerStarted","Data":"4791843c23c9d6750e6cb0f977f6b7833929766197580316c042766b872e3509"} Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.320497 4934 generic.go:334] "Generic (PLEG): container finished" podID="2ee60051-8eec-4da3-ae78-b409c5cdf9f3" containerID="52c0e29ee5516233875492f1f454adace2d7af3770ceda4458f1fdec613b5758" exitCode=0 Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.320580 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtq22" event={"ID":"2ee60051-8eec-4da3-ae78-b409c5cdf9f3","Type":"ContainerDied","Data":"52c0e29ee5516233875492f1f454adace2d7af3770ceda4458f1fdec613b5758"} Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.320644 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtq22" event={"ID":"2ee60051-8eec-4da3-ae78-b409c5cdf9f3","Type":"ContainerStarted","Data":"1eaa4097b44192e81d6544f28960aaed17e2a122a5a26973ed8557daf0c2b248"} Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.449666 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sz4sq"] Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.451254 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sz4sq" Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.455982 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.474739 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sz4sq"] Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.494903 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgxln\" (UniqueName: \"kubernetes.io/projected/72569b30-4bc6-4f58-b62e-c1a0ef0961d6-kube-api-access-rgxln\") pod \"community-operators-sz4sq\" (UID: \"72569b30-4bc6-4f58-b62e-c1a0ef0961d6\") " pod="openshift-marketplace/community-operators-sz4sq" Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.495023 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72569b30-4bc6-4f58-b62e-c1a0ef0961d6-utilities\") pod \"community-operators-sz4sq\" (UID: \"72569b30-4bc6-4f58-b62e-c1a0ef0961d6\") " pod="openshift-marketplace/community-operators-sz4sq" Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.495128 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72569b30-4bc6-4f58-b62e-c1a0ef0961d6-catalog-content\") pod \"community-operators-sz4sq\" (UID: \"72569b30-4bc6-4f58-b62e-c1a0ef0961d6\") " pod="openshift-marketplace/community-operators-sz4sq" Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.596842 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72569b30-4bc6-4f58-b62e-c1a0ef0961d6-utilities\") pod \"community-operators-sz4sq\" (UID: \"72569b30-4bc6-4f58-b62e-c1a0ef0961d6\") " pod="openshift-marketplace/community-operators-sz4sq" Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.596939 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72569b30-4bc6-4f58-b62e-c1a0ef0961d6-catalog-content\") pod \"community-operators-sz4sq\" (UID: \"72569b30-4bc6-4f58-b62e-c1a0ef0961d6\") " pod="openshift-marketplace/community-operators-sz4sq" Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.596984 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgxln\" (UniqueName: \"kubernetes.io/projected/72569b30-4bc6-4f58-b62e-c1a0ef0961d6-kube-api-access-rgxln\") pod \"community-operators-sz4sq\" (UID: \"72569b30-4bc6-4f58-b62e-c1a0ef0961d6\") " pod="openshift-marketplace/community-operators-sz4sq" Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.597784 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72569b30-4bc6-4f58-b62e-c1a0ef0961d6-utilities\") pod \"community-operators-sz4sq\" (UID: \"72569b30-4bc6-4f58-b62e-c1a0ef0961d6\") " pod="openshift-marketplace/community-operators-sz4sq" Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.598074 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72569b30-4bc6-4f58-b62e-c1a0ef0961d6-catalog-content\") pod \"community-operators-sz4sq\" (UID: \"72569b30-4bc6-4f58-b62e-c1a0ef0961d6\") " pod="openshift-marketplace/community-operators-sz4sq" Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.618816 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgxln\" (UniqueName: \"kubernetes.io/projected/72569b30-4bc6-4f58-b62e-c1a0ef0961d6-kube-api-access-rgxln\") pod \"community-operators-sz4sq\" (UID: \"72569b30-4bc6-4f58-b62e-c1a0ef0961d6\") " pod="openshift-marketplace/community-operators-sz4sq" Dec 27 07:48:38 crc kubenswrapper[4934]: I1227 07:48:38.770985 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sz4sq" Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.051616 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7rs6w"] Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.053218 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7rs6w" Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.055606 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.060135 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7rs6w"] Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.103595 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3-catalog-content\") pod \"redhat-marketplace-7rs6w\" (UID: \"ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3\") " pod="openshift-marketplace/redhat-marketplace-7rs6w" Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.103659 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3-utilities\") pod \"redhat-marketplace-7rs6w\" (UID: \"ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3\") " pod="openshift-marketplace/redhat-marketplace-7rs6w" Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.103721 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw5ms\" (UniqueName: \"kubernetes.io/projected/ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3-kube-api-access-gw5ms\") pod \"redhat-marketplace-7rs6w\" (UID: \"ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3\") " pod="openshift-marketplace/redhat-marketplace-7rs6w" Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.204717 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3-utilities\") pod \"redhat-marketplace-7rs6w\" (UID: \"ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3\") " pod="openshift-marketplace/redhat-marketplace-7rs6w" Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.204895 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw5ms\" (UniqueName: \"kubernetes.io/projected/ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3-kube-api-access-gw5ms\") pod \"redhat-marketplace-7rs6w\" (UID: \"ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3\") " pod="openshift-marketplace/redhat-marketplace-7rs6w" Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.204945 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3-catalog-content\") pod \"redhat-marketplace-7rs6w\" (UID: \"ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3\") " pod="openshift-marketplace/redhat-marketplace-7rs6w" Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.206067 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3-utilities\") pod \"redhat-marketplace-7rs6w\" (UID: \"ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3\") " pod="openshift-marketplace/redhat-marketplace-7rs6w" Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.206151 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3-catalog-content\") pod \"redhat-marketplace-7rs6w\" (UID: \"ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3\") " pod="openshift-marketplace/redhat-marketplace-7rs6w" Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.235847 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sz4sq"] Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.242280 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw5ms\" (UniqueName: \"kubernetes.io/projected/ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3-kube-api-access-gw5ms\") pod \"redhat-marketplace-7rs6w\" (UID: \"ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3\") " pod="openshift-marketplace/redhat-marketplace-7rs6w" Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.291339 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.291388 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.297483 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.329663 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_40c7cd23-14a7-4b30-9639-84d83ba74fe3/prometheus/1.log" Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.333733 4934 scope.go:117] "RemoveContainer" containerID="41c879bc517aa226186963efe3788cca745786da2466030eb60387789ec75d06" Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.334014 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sz4sq" event={"ID":"72569b30-4bc6-4f58-b62e-c1a0ef0961d6","Type":"ContainerStarted","Data":"f1b833b1651b97a4fab56c4f2d1adfa203b84fbcd2c60b024fe588692b0b434c"} Dec 27 07:48:39 crc kubenswrapper[4934]: E1227 07:48:39.334580 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=prometheus pod=prometheus-k8s-0_openshift-monitoring(40c7cd23-14a7-4b30-9639-84d83ba74fe3)\"" pod="openshift-monitoring/prometheus-k8s-0" podUID="40c7cd23-14a7-4b30-9639-84d83ba74fe3" Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.341809 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.390165 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7rs6w" Dec 27 07:48:39 crc kubenswrapper[4934]: I1227 07:48:39.419339 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-p8c2x"] Dec 27 07:48:40 crc kubenswrapper[4934]: I1227 07:48:40.222907 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7rs6w"] Dec 27 07:48:40 crc kubenswrapper[4934]: W1227 07:48:40.238326 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec69ef3c_f87e_4b0d_ba44_9067f8e4a2b3.slice/crio-b68746d97e09f96a3a220dd5adbe67265a5e6edfc4b30934a9b9d3c08d431865 WatchSource:0}: Error finding container b68746d97e09f96a3a220dd5adbe67265a5e6edfc4b30934a9b9d3c08d431865: Status 404 returned error can't find the container with id b68746d97e09f96a3a220dd5adbe67265a5e6edfc4b30934a9b9d3c08d431865 Dec 27 07:48:40 crc kubenswrapper[4934]: I1227 07:48:40.344980 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtq22" event={"ID":"2ee60051-8eec-4da3-ae78-b409c5cdf9f3","Type":"ContainerStarted","Data":"4ed55487100866c88ecfd2902bb50041fbdd36462df8faef2de5af3edae71443"} Dec 27 07:48:40 crc kubenswrapper[4934]: I1227 07:48:40.347761 4934 generic.go:334] "Generic (PLEG): container finished" podID="0661ef0e-af9d-41ba-8059-2e878b11b042" containerID="1f02dd9201c7431ad29f848d25bbce303a821f8e23b9a5c024d34e3007f71ca7" exitCode=0 Dec 27 07:48:40 crc kubenswrapper[4934]: I1227 07:48:40.347850 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lbfd2" event={"ID":"0661ef0e-af9d-41ba-8059-2e878b11b042","Type":"ContainerDied","Data":"1f02dd9201c7431ad29f848d25bbce303a821f8e23b9a5c024d34e3007f71ca7"} Dec 27 07:48:40 crc kubenswrapper[4934]: I1227 07:48:40.349991 4934 generic.go:334] "Generic (PLEG): container finished" podID="72569b30-4bc6-4f58-b62e-c1a0ef0961d6" containerID="e97a8bb3634697de5c79811fe759016ec41f1e11f97f96966aeaaa5cfa8bc9e5" exitCode=0 Dec 27 07:48:40 crc kubenswrapper[4934]: I1227 07:48:40.350026 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sz4sq" event={"ID":"72569b30-4bc6-4f58-b62e-c1a0ef0961d6","Type":"ContainerDied","Data":"e97a8bb3634697de5c79811fe759016ec41f1e11f97f96966aeaaa5cfa8bc9e5"} Dec 27 07:48:40 crc kubenswrapper[4934]: I1227 07:48:40.352240 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7rs6w" event={"ID":"ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3","Type":"ContainerStarted","Data":"b68746d97e09f96a3a220dd5adbe67265a5e6edfc4b30934a9b9d3c08d431865"} Dec 27 07:48:40 crc kubenswrapper[4934]: I1227 07:48:40.952337 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:40 crc kubenswrapper[4934]: I1227 07:48:40.952825 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:48:40 crc kubenswrapper[4934]: I1227 07:48:40.953702 4934 scope.go:117] "RemoveContainer" containerID="41c879bc517aa226186963efe3788cca745786da2466030eb60387789ec75d06" Dec 27 07:48:40 crc kubenswrapper[4934]: E1227 07:48:40.954287 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=prometheus pod=prometheus-k8s-0_openshift-monitoring(40c7cd23-14a7-4b30-9639-84d83ba74fe3)\"" pod="openshift-monitoring/prometheus-k8s-0" podUID="40c7cd23-14a7-4b30-9639-84d83ba74fe3" Dec 27 07:48:41 crc kubenswrapper[4934]: I1227 07:48:41.360478 4934 generic.go:334] "Generic (PLEG): container finished" podID="ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3" containerID="4ff160d17e909e8345091a08d59416f2cf106c56f13ea5c1c24ef322cb8c4e8c" exitCode=0 Dec 27 07:48:41 crc kubenswrapper[4934]: I1227 07:48:41.360595 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7rs6w" event={"ID":"ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3","Type":"ContainerDied","Data":"4ff160d17e909e8345091a08d59416f2cf106c56f13ea5c1c24ef322cb8c4e8c"} Dec 27 07:48:41 crc kubenswrapper[4934]: I1227 07:48:41.363335 4934 generic.go:334] "Generic (PLEG): container finished" podID="2ee60051-8eec-4da3-ae78-b409c5cdf9f3" containerID="4ed55487100866c88ecfd2902bb50041fbdd36462df8faef2de5af3edae71443" exitCode=0 Dec 27 07:48:41 crc kubenswrapper[4934]: I1227 07:48:41.363407 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtq22" event={"ID":"2ee60051-8eec-4da3-ae78-b409c5cdf9f3","Type":"ContainerDied","Data":"4ed55487100866c88ecfd2902bb50041fbdd36462df8faef2de5af3edae71443"} Dec 27 07:48:41 crc kubenswrapper[4934]: I1227 07:48:41.366612 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lbfd2" event={"ID":"0661ef0e-af9d-41ba-8059-2e878b11b042","Type":"ContainerStarted","Data":"89e9d67eb5875cd1abfba344665069bf1eb778ea3464c19c26c91310ae75f4bb"} Dec 27 07:48:41 crc kubenswrapper[4934]: I1227 07:48:41.403596 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lbfd2" podStartSLOduration=2.812708162 podStartE2EDuration="5.403573421s" podCreationTimestamp="2025-12-27 07:48:36 +0000 UTC" firstStartedPulling="2025-12-27 07:48:38.310869303 +0000 UTC m=+379.131309897" lastFinishedPulling="2025-12-27 07:48:40.901734562 +0000 UTC m=+381.722175156" observedRunningTime="2025-12-27 07:48:41.401401093 +0000 UTC m=+382.221841697" watchObservedRunningTime="2025-12-27 07:48:41.403573421 +0000 UTC m=+382.224014015" Dec 27 07:48:42 crc kubenswrapper[4934]: I1227 07:48:42.374152 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtq22" event={"ID":"2ee60051-8eec-4da3-ae78-b409c5cdf9f3","Type":"ContainerStarted","Data":"15e9661eb330205503e9cf73bf9f1ef6051450428bd4903c17aa161894e106f8"} Dec 27 07:48:42 crc kubenswrapper[4934]: I1227 07:48:42.375568 4934 generic.go:334] "Generic (PLEG): container finished" podID="72569b30-4bc6-4f58-b62e-c1a0ef0961d6" containerID="c474f93335f5188402eb866ebd1992166708cac0f42b9b122806a2a61583c872" exitCode=0 Dec 27 07:48:42 crc kubenswrapper[4934]: I1227 07:48:42.375626 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sz4sq" event={"ID":"72569b30-4bc6-4f58-b62e-c1a0ef0961d6","Type":"ContainerDied","Data":"c474f93335f5188402eb866ebd1992166708cac0f42b9b122806a2a61583c872"} Dec 27 07:48:42 crc kubenswrapper[4934]: I1227 07:48:42.377914 4934 generic.go:334] "Generic (PLEG): container finished" podID="ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3" containerID="92773a3214bd94fe79b7b5c8e5287c67a7de71ea28be4ca465f21b93ff6bbed4" exitCode=0 Dec 27 07:48:42 crc kubenswrapper[4934]: I1227 07:48:42.377978 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7rs6w" event={"ID":"ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3","Type":"ContainerDied","Data":"92773a3214bd94fe79b7b5c8e5287c67a7de71ea28be4ca465f21b93ff6bbed4"} Dec 27 07:48:42 crc kubenswrapper[4934]: I1227 07:48:42.400017 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qtq22" podStartSLOduration=2.958246375 podStartE2EDuration="6.399998279s" podCreationTimestamp="2025-12-27 07:48:36 +0000 UTC" firstStartedPulling="2025-12-27 07:48:38.323241206 +0000 UTC m=+379.143681800" lastFinishedPulling="2025-12-27 07:48:41.76499311 +0000 UTC m=+382.585433704" observedRunningTime="2025-12-27 07:48:42.393858804 +0000 UTC m=+383.214299398" watchObservedRunningTime="2025-12-27 07:48:42.399998279 +0000 UTC m=+383.220438873" Dec 27 07:48:43 crc kubenswrapper[4934]: I1227 07:48:43.386149 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sz4sq" event={"ID":"72569b30-4bc6-4f58-b62e-c1a0ef0961d6","Type":"ContainerStarted","Data":"7c687fb3753437c03baff26d416f53d218c8d9e378449a9edf53e76c4acb0966"} Dec 27 07:48:43 crc kubenswrapper[4934]: I1227 07:48:43.388343 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7rs6w" event={"ID":"ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3","Type":"ContainerStarted","Data":"07946dca35da56c54729d4d20d3bf767ec79ff82cec1df4f7a339d5a75a98d2c"} Dec 27 07:48:43 crc kubenswrapper[4934]: I1227 07:48:43.421583 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sz4sq" podStartSLOduration=2.961529139 podStartE2EDuration="5.421564981s" podCreationTimestamp="2025-12-27 07:48:38 +0000 UTC" firstStartedPulling="2025-12-27 07:48:40.35136278 +0000 UTC m=+381.171803374" lastFinishedPulling="2025-12-27 07:48:42.811398622 +0000 UTC m=+383.631839216" observedRunningTime="2025-12-27 07:48:43.407878489 +0000 UTC m=+384.228319113" watchObservedRunningTime="2025-12-27 07:48:43.421564981 +0000 UTC m=+384.242005565" Dec 27 07:48:43 crc kubenswrapper[4934]: I1227 07:48:43.449131 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7rs6w" podStartSLOduration=3.052102347 podStartE2EDuration="4.449107189s" podCreationTimestamp="2025-12-27 07:48:39 +0000 UTC" firstStartedPulling="2025-12-27 07:48:41.36275387 +0000 UTC m=+382.183194474" lastFinishedPulling="2025-12-27 07:48:42.759758712 +0000 UTC m=+383.580199316" observedRunningTime="2025-12-27 07:48:43.437115665 +0000 UTC m=+384.257556269" watchObservedRunningTime="2025-12-27 07:48:43.449107189 +0000 UTC m=+384.269547783" Dec 27 07:48:45 crc kubenswrapper[4934]: I1227 07:48:45.330041 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 07:48:45 crc kubenswrapper[4934]: I1227 07:48:45.330479 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 07:48:46 crc kubenswrapper[4934]: I1227 07:48:46.391032 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qtq22" Dec 27 07:48:46 crc kubenswrapper[4934]: I1227 07:48:46.393888 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qtq22" Dec 27 07:48:47 crc kubenswrapper[4934]: I1227 07:48:47.006968 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lbfd2" Dec 27 07:48:47 crc kubenswrapper[4934]: I1227 07:48:47.007467 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lbfd2" Dec 27 07:48:47 crc kubenswrapper[4934]: I1227 07:48:47.058215 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lbfd2" Dec 27 07:48:47 crc kubenswrapper[4934]: I1227 07:48:47.457912 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lbfd2" Dec 27 07:48:47 crc kubenswrapper[4934]: I1227 07:48:47.460027 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qtq22" podUID="2ee60051-8eec-4da3-ae78-b409c5cdf9f3" containerName="registry-server" probeResult="failure" output=< Dec 27 07:48:47 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 07:48:47 crc kubenswrapper[4934]: > Dec 27 07:48:48 crc kubenswrapper[4934]: I1227 07:48:48.771946 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sz4sq" Dec 27 07:48:48 crc kubenswrapper[4934]: I1227 07:48:48.771995 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sz4sq" Dec 27 07:48:48 crc kubenswrapper[4934]: I1227 07:48:48.819579 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sz4sq" Dec 27 07:48:49 crc kubenswrapper[4934]: I1227 07:48:49.390771 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7rs6w" Dec 27 07:48:49 crc kubenswrapper[4934]: I1227 07:48:49.391125 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7rs6w" Dec 27 07:48:49 crc kubenswrapper[4934]: I1227 07:48:49.449977 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7rs6w" Dec 27 07:48:49 crc kubenswrapper[4934]: I1227 07:48:49.487606 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sz4sq" Dec 27 07:48:49 crc kubenswrapper[4934]: I1227 07:48:49.528817 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7rs6w" Dec 27 07:48:49 crc kubenswrapper[4934]: I1227 07:48:49.739779 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:49 crc kubenswrapper[4934]: I1227 07:48:49.739827 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:48:49 crc kubenswrapper[4934]: I1227 07:48:49.751832 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-589f9f9656-pbctv"] Dec 27 07:48:49 crc kubenswrapper[4934]: I1227 07:48:49.752047 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" podUID="46338877-dff9-4023-b399-212fce6fae21" containerName="controller-manager" containerID="cri-o://bad748b3be090ee97861e441494efe15616c155cc762df3c3fad147ecb21ed08" gracePeriod=30 Dec 27 07:48:52 crc kubenswrapper[4934]: I1227 07:48:52.467961 4934 scope.go:117] "RemoveContainer" containerID="41c879bc517aa226186963efe3788cca745786da2466030eb60387789ec75d06" Dec 27 07:48:53 crc kubenswrapper[4934]: I1227 07:48:53.180687 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" Dec 27 07:48:53 crc kubenswrapper[4934]: I1227 07:48:53.288456 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ln99b"] Dec 27 07:48:54 crc kubenswrapper[4934]: I1227 07:48:54.005937 4934 patch_prober.go:28] interesting pod/controller-manager-589f9f9656-pbctv container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.64:8443/healthz\": dial tcp 10.217.0.64:8443: connect: connection refused" start-of-body= Dec 27 07:48:54 crc kubenswrapper[4934]: I1227 07:48:54.006022 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" podUID="46338877-dff9-4023-b399-212fce6fae21" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.64:8443/healthz\": dial tcp 10.217.0.64:8443: connect: connection refused" Dec 27 07:48:56 crc kubenswrapper[4934]: I1227 07:48:56.448886 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qtq22" Dec 27 07:48:56 crc kubenswrapper[4934]: I1227 07:48:56.509266 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qtq22" Dec 27 07:48:59 crc kubenswrapper[4934]: I1227 07:48:59.709506 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-589f9f9656-pbctv_46338877-dff9-4023-b399-212fce6fae21/controller-manager/0.log" Dec 27 07:48:59 crc kubenswrapper[4934]: I1227 07:48:59.709898 4934 generic.go:334] "Generic (PLEG): container finished" podID="46338877-dff9-4023-b399-212fce6fae21" containerID="bad748b3be090ee97861e441494efe15616c155cc762df3c3fad147ecb21ed08" exitCode=-1 Dec 27 07:48:59 crc kubenswrapper[4934]: I1227 07:48:59.710076 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" event={"ID":"46338877-dff9-4023-b399-212fce6fae21","Type":"ContainerDied","Data":"bad748b3be090ee97861e441494efe15616c155cc762df3c3fad147ecb21ed08"} Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.118190 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.154623 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7967777f79-kzj4d"] Dec 27 07:49:00 crc kubenswrapper[4934]: E1227 07:49:00.154951 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46338877-dff9-4023-b399-212fce6fae21" containerName="controller-manager" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.154972 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="46338877-dff9-4023-b399-212fce6fae21" containerName="controller-manager" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.155118 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="46338877-dff9-4023-b399-212fce6fae21" containerName="controller-manager" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.155625 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.165963 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7967777f79-kzj4d"] Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.300261 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47vjv\" (UniqueName: \"kubernetes.io/projected/46338877-dff9-4023-b399-212fce6fae21-kube-api-access-47vjv\") pod \"46338877-dff9-4023-b399-212fce6fae21\" (UID: \"46338877-dff9-4023-b399-212fce6fae21\") " Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.300361 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/46338877-dff9-4023-b399-212fce6fae21-client-ca\") pod \"46338877-dff9-4023-b399-212fce6fae21\" (UID: \"46338877-dff9-4023-b399-212fce6fae21\") " Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.300417 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46338877-dff9-4023-b399-212fce6fae21-serving-cert\") pod \"46338877-dff9-4023-b399-212fce6fae21\" (UID: \"46338877-dff9-4023-b399-212fce6fae21\") " Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.300447 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/46338877-dff9-4023-b399-212fce6fae21-proxy-ca-bundles\") pod \"46338877-dff9-4023-b399-212fce6fae21\" (UID: \"46338877-dff9-4023-b399-212fce6fae21\") " Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.300509 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46338877-dff9-4023-b399-212fce6fae21-config\") pod \"46338877-dff9-4023-b399-212fce6fae21\" (UID: \"46338877-dff9-4023-b399-212fce6fae21\") " Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.300685 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2a688ece-0e00-4241-b7dd-9dde25f0eaf1-client-ca\") pod \"controller-manager-7967777f79-kzj4d\" (UID: \"2a688ece-0e00-4241-b7dd-9dde25f0eaf1\") " pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.300733 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2a688ece-0e00-4241-b7dd-9dde25f0eaf1-proxy-ca-bundles\") pod \"controller-manager-7967777f79-kzj4d\" (UID: \"2a688ece-0e00-4241-b7dd-9dde25f0eaf1\") " pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.300779 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brcdx\" (UniqueName: \"kubernetes.io/projected/2a688ece-0e00-4241-b7dd-9dde25f0eaf1-kube-api-access-brcdx\") pod \"controller-manager-7967777f79-kzj4d\" (UID: \"2a688ece-0e00-4241-b7dd-9dde25f0eaf1\") " pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.300824 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a688ece-0e00-4241-b7dd-9dde25f0eaf1-config\") pod \"controller-manager-7967777f79-kzj4d\" (UID: \"2a688ece-0e00-4241-b7dd-9dde25f0eaf1\") " pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.300852 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a688ece-0e00-4241-b7dd-9dde25f0eaf1-serving-cert\") pod \"controller-manager-7967777f79-kzj4d\" (UID: \"2a688ece-0e00-4241-b7dd-9dde25f0eaf1\") " pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.301858 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46338877-dff9-4023-b399-212fce6fae21-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "46338877-dff9-4023-b399-212fce6fae21" (UID: "46338877-dff9-4023-b399-212fce6fae21"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.302241 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46338877-dff9-4023-b399-212fce6fae21-config" (OuterVolumeSpecName: "config") pod "46338877-dff9-4023-b399-212fce6fae21" (UID: "46338877-dff9-4023-b399-212fce6fae21"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.302408 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46338877-dff9-4023-b399-212fce6fae21-client-ca" (OuterVolumeSpecName: "client-ca") pod "46338877-dff9-4023-b399-212fce6fae21" (UID: "46338877-dff9-4023-b399-212fce6fae21"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.310011 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46338877-dff9-4023-b399-212fce6fae21-kube-api-access-47vjv" (OuterVolumeSpecName: "kube-api-access-47vjv") pod "46338877-dff9-4023-b399-212fce6fae21" (UID: "46338877-dff9-4023-b399-212fce6fae21"). InnerVolumeSpecName "kube-api-access-47vjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.318300 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46338877-dff9-4023-b399-212fce6fae21-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "46338877-dff9-4023-b399-212fce6fae21" (UID: "46338877-dff9-4023-b399-212fce6fae21"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.403225 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brcdx\" (UniqueName: \"kubernetes.io/projected/2a688ece-0e00-4241-b7dd-9dde25f0eaf1-kube-api-access-brcdx\") pod \"controller-manager-7967777f79-kzj4d\" (UID: \"2a688ece-0e00-4241-b7dd-9dde25f0eaf1\") " pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.403335 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a688ece-0e00-4241-b7dd-9dde25f0eaf1-config\") pod \"controller-manager-7967777f79-kzj4d\" (UID: \"2a688ece-0e00-4241-b7dd-9dde25f0eaf1\") " pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.403371 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a688ece-0e00-4241-b7dd-9dde25f0eaf1-serving-cert\") pod \"controller-manager-7967777f79-kzj4d\" (UID: \"2a688ece-0e00-4241-b7dd-9dde25f0eaf1\") " pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.403399 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2a688ece-0e00-4241-b7dd-9dde25f0eaf1-client-ca\") pod \"controller-manager-7967777f79-kzj4d\" (UID: \"2a688ece-0e00-4241-b7dd-9dde25f0eaf1\") " pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.403449 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2a688ece-0e00-4241-b7dd-9dde25f0eaf1-proxy-ca-bundles\") pod \"controller-manager-7967777f79-kzj4d\" (UID: \"2a688ece-0e00-4241-b7dd-9dde25f0eaf1\") " pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.403510 4934 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/46338877-dff9-4023-b399-212fce6fae21-client-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.403525 4934 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46338877-dff9-4023-b399-212fce6fae21-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.403538 4934 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/46338877-dff9-4023-b399-212fce6fae21-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.403552 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46338877-dff9-4023-b399-212fce6fae21-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.403563 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47vjv\" (UniqueName: \"kubernetes.io/projected/46338877-dff9-4023-b399-212fce6fae21-kube-api-access-47vjv\") on node \"crc\" DevicePath \"\"" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.404751 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2a688ece-0e00-4241-b7dd-9dde25f0eaf1-proxy-ca-bundles\") pod \"controller-manager-7967777f79-kzj4d\" (UID: \"2a688ece-0e00-4241-b7dd-9dde25f0eaf1\") " pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.404964 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a688ece-0e00-4241-b7dd-9dde25f0eaf1-config\") pod \"controller-manager-7967777f79-kzj4d\" (UID: \"2a688ece-0e00-4241-b7dd-9dde25f0eaf1\") " pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.405567 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2a688ece-0e00-4241-b7dd-9dde25f0eaf1-client-ca\") pod \"controller-manager-7967777f79-kzj4d\" (UID: \"2a688ece-0e00-4241-b7dd-9dde25f0eaf1\") " pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.410708 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2a688ece-0e00-4241-b7dd-9dde25f0eaf1-serving-cert\") pod \"controller-manager-7967777f79-kzj4d\" (UID: \"2a688ece-0e00-4241-b7dd-9dde25f0eaf1\") " pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.420936 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brcdx\" (UniqueName: \"kubernetes.io/projected/2a688ece-0e00-4241-b7dd-9dde25f0eaf1-kube-api-access-brcdx\") pod \"controller-manager-7967777f79-kzj4d\" (UID: \"2a688ece-0e00-4241-b7dd-9dde25f0eaf1\") " pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.479354 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.720351 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" event={"ID":"46338877-dff9-4023-b399-212fce6fae21","Type":"ContainerDied","Data":"3059cdf7a39b73ce325003d215e8d003ac73ffa9603987f2fe77af21ace3b80e"} Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.720389 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-589f9f9656-pbctv" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.720426 4934 scope.go:117] "RemoveContainer" containerID="bad748b3be090ee97861e441494efe15616c155cc762df3c3fad147ecb21ed08" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.723174 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_40c7cd23-14a7-4b30-9639-84d83ba74fe3/prometheus/2.log" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.724191 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_40c7cd23-14a7-4b30-9639-84d83ba74fe3/prometheus/1.log" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.726250 4934 generic.go:334] "Generic (PLEG): container finished" podID="40c7cd23-14a7-4b30-9639-84d83ba74fe3" containerID="fd35b2027f2556f022c6c16a393209f56220481bc406e1e57f5f9ac3f4134dfe" exitCode=1 Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.726349 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"40c7cd23-14a7-4b30-9639-84d83ba74fe3","Type":"ContainerDied","Data":"fd35b2027f2556f022c6c16a393209f56220481bc406e1e57f5f9ac3f4134dfe"} Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.727159 4934 scope.go:117] "RemoveContainer" containerID="fd35b2027f2556f022c6c16a393209f56220481bc406e1e57f5f9ac3f4134dfe" Dec 27 07:49:00 crc kubenswrapper[4934]: E1227 07:49:00.727619 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=prometheus pod=prometheus-k8s-0_openshift-monitoring(40c7cd23-14a7-4b30-9639-84d83ba74fe3)\"" pod="openshift-monitoring/prometheus-k8s-0" podUID="40c7cd23-14a7-4b30-9639-84d83ba74fe3" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.740620 4934 scope.go:117] "RemoveContainer" containerID="41c879bc517aa226186963efe3788cca745786da2466030eb60387789ec75d06" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.775578 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-589f9f9656-pbctv"] Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.781430 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-589f9f9656-pbctv"] Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.937067 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7967777f79-kzj4d"] Dec 27 07:49:00 crc kubenswrapper[4934]: W1227 07:49:00.941268 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a688ece_0e00_4241_b7dd_9dde25f0eaf1.slice/crio-7fd8f3b43eb51300d71e9d5c69715b21d101f4a107d3165489d5b3cdcedfc0fb WatchSource:0}: Error finding container 7fd8f3b43eb51300d71e9d5c69715b21d101f4a107d3165489d5b3cdcedfc0fb: Status 404 returned error can't find the container with id 7fd8f3b43eb51300d71e9d5c69715b21d101f4a107d3165489d5b3cdcedfc0fb Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.953004 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:49:00 crc kubenswrapper[4934]: I1227 07:49:00.953067 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:49:01 crc kubenswrapper[4934]: I1227 07:49:01.474516 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46338877-dff9-4023-b399-212fce6fae21" path="/var/lib/kubelet/pods/46338877-dff9-4023-b399-212fce6fae21/volumes" Dec 27 07:49:01 crc kubenswrapper[4934]: I1227 07:49:01.733940 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_40c7cd23-14a7-4b30-9639-84d83ba74fe3/prometheus/2.log" Dec 27 07:49:01 crc kubenswrapper[4934]: I1227 07:49:01.736592 4934 scope.go:117] "RemoveContainer" containerID="fd35b2027f2556f022c6c16a393209f56220481bc406e1e57f5f9ac3f4134dfe" Dec 27 07:49:01 crc kubenswrapper[4934]: E1227 07:49:01.737065 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=prometheus pod=prometheus-k8s-0_openshift-monitoring(40c7cd23-14a7-4b30-9639-84d83ba74fe3)\"" pod="openshift-monitoring/prometheus-k8s-0" podUID="40c7cd23-14a7-4b30-9639-84d83ba74fe3" Dec 27 07:49:01 crc kubenswrapper[4934]: I1227 07:49:01.738781 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" event={"ID":"2a688ece-0e00-4241-b7dd-9dde25f0eaf1","Type":"ContainerStarted","Data":"fd15bc9e9543c8afdb985cb0426e5686fff142c76fd4188ddea84d8885e949cc"} Dec 27 07:49:01 crc kubenswrapper[4934]: I1227 07:49:01.738823 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" event={"ID":"2a688ece-0e00-4241-b7dd-9dde25f0eaf1","Type":"ContainerStarted","Data":"7fd8f3b43eb51300d71e9d5c69715b21d101f4a107d3165489d5b3cdcedfc0fb"} Dec 27 07:49:01 crc kubenswrapper[4934]: I1227 07:49:01.739594 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 07:49:01 crc kubenswrapper[4934]: I1227 07:49:01.743899 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 07:49:01 crc kubenswrapper[4934]: I1227 07:49:01.833854 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" podStartSLOduration=12.833836162 podStartE2EDuration="12.833836162s" podCreationTimestamp="2025-12-27 07:48:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:49:01.786894656 +0000 UTC m=+402.607335270" watchObservedRunningTime="2025-12-27 07:49:01.833836162 +0000 UTC m=+402.654276756" Dec 27 07:49:04 crc kubenswrapper[4934]: I1227 07:49:04.466113 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-p8c2x" podUID="091c02c2-ae22-44e8-bf6b-64b69d6d9d83" containerName="console" containerID="cri-o://109e74a00ca3e09f9af2206b0495f03664b3ca731e65fe55e0b9e8066f0d23a3" gracePeriod=15 Dec 27 07:49:04 crc kubenswrapper[4934]: E1227 07:49:04.631257 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod091c02c2_ae22_44e8_bf6b_64b69d6d9d83.slice/crio-conmon-109e74a00ca3e09f9af2206b0495f03664b3ca731e65fe55e0b9e8066f0d23a3.scope\": RecentStats: unable to find data in memory cache]" Dec 27 07:49:04 crc kubenswrapper[4934]: I1227 07:49:04.764765 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-p8c2x_091c02c2-ae22-44e8-bf6b-64b69d6d9d83/console/0.log" Dec 27 07:49:04 crc kubenswrapper[4934]: I1227 07:49:04.764821 4934 generic.go:334] "Generic (PLEG): container finished" podID="091c02c2-ae22-44e8-bf6b-64b69d6d9d83" containerID="109e74a00ca3e09f9af2206b0495f03664b3ca731e65fe55e0b9e8066f0d23a3" exitCode=2 Dec 27 07:49:04 crc kubenswrapper[4934]: I1227 07:49:04.764854 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-p8c2x" event={"ID":"091c02c2-ae22-44e8-bf6b-64b69d6d9d83","Type":"ContainerDied","Data":"109e74a00ca3e09f9af2206b0495f03664b3ca731e65fe55e0b9e8066f0d23a3"} Dec 27 07:49:04 crc kubenswrapper[4934]: I1227 07:49:04.899588 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-p8c2x_091c02c2-ae22-44e8-bf6b-64b69d6d9d83/console/0.log" Dec 27 07:49:04 crc kubenswrapper[4934]: I1227 07:49:04.899649 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.091833 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-service-ca\") pod \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.091908 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-trusted-ca-bundle\") pod \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.091935 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-console-serving-cert\") pod \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.091970 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-console-oauth-config\") pod \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.092081 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78dm8\" (UniqueName: \"kubernetes.io/projected/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-kube-api-access-78dm8\") pod \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.092173 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-console-config\") pod \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.092200 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-oauth-serving-cert\") pod \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\" (UID: \"091c02c2-ae22-44e8-bf6b-64b69d6d9d83\") " Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.093075 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "091c02c2-ae22-44e8-bf6b-64b69d6d9d83" (UID: "091c02c2-ae22-44e8-bf6b-64b69d6d9d83"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.093059 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-service-ca" (OuterVolumeSpecName: "service-ca") pod "091c02c2-ae22-44e8-bf6b-64b69d6d9d83" (UID: "091c02c2-ae22-44e8-bf6b-64b69d6d9d83"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.093801 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-console-config" (OuterVolumeSpecName: "console-config") pod "091c02c2-ae22-44e8-bf6b-64b69d6d9d83" (UID: "091c02c2-ae22-44e8-bf6b-64b69d6d9d83"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.094291 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "091c02c2-ae22-44e8-bf6b-64b69d6d9d83" (UID: "091c02c2-ae22-44e8-bf6b-64b69d6d9d83"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.098579 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "091c02c2-ae22-44e8-bf6b-64b69d6d9d83" (UID: "091c02c2-ae22-44e8-bf6b-64b69d6d9d83"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.098902 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-kube-api-access-78dm8" (OuterVolumeSpecName: "kube-api-access-78dm8") pod "091c02c2-ae22-44e8-bf6b-64b69d6d9d83" (UID: "091c02c2-ae22-44e8-bf6b-64b69d6d9d83"). InnerVolumeSpecName "kube-api-access-78dm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.098931 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "091c02c2-ae22-44e8-bf6b-64b69d6d9d83" (UID: "091c02c2-ae22-44e8-bf6b-64b69d6d9d83"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.193713 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78dm8\" (UniqueName: \"kubernetes.io/projected/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-kube-api-access-78dm8\") on node \"crc\" DevicePath \"\"" Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.193757 4934 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-console-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.193766 4934 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.193775 4934 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-service-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.193785 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.193792 4934 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.193800 4934 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/091c02c2-ae22-44e8-bf6b-64b69d6d9d83-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.772422 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-p8c2x_091c02c2-ae22-44e8-bf6b-64b69d6d9d83/console/0.log" Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.772493 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-p8c2x" event={"ID":"091c02c2-ae22-44e8-bf6b-64b69d6d9d83","Type":"ContainerDied","Data":"947b1aad9814ff50655460b476d68e661d3b2a458cec61bdef9671bd855af5dd"} Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.772545 4934 scope.go:117] "RemoveContainer" containerID="109e74a00ca3e09f9af2206b0495f03664b3ca731e65fe55e0b9e8066f0d23a3" Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.772547 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-p8c2x" Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.799054 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-p8c2x"] Dec 27 07:49:05 crc kubenswrapper[4934]: I1227 07:49:05.804443 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-p8c2x"] Dec 27 07:49:07 crc kubenswrapper[4934]: I1227 07:49:07.476847 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="091c02c2-ae22-44e8-bf6b-64b69d6d9d83" path="/var/lib/kubelet/pods/091c02c2-ae22-44e8-bf6b-64b69d6d9d83/volumes" Dec 27 07:49:09 crc kubenswrapper[4934]: I1227 07:49:09.747160 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:49:09 crc kubenswrapper[4934]: I1227 07:49:09.765036 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" Dec 27 07:49:15 crc kubenswrapper[4934]: I1227 07:49:15.329963 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 07:49:15 crc kubenswrapper[4934]: I1227 07:49:15.330421 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 07:49:15 crc kubenswrapper[4934]: I1227 07:49:15.330489 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 07:49:15 crc kubenswrapper[4934]: I1227 07:49:15.331509 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"44124e136c1edec767fa7b55e82033376dbd4c05224d1c80d3b03b40d50a5a60"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 07:49:15 crc kubenswrapper[4934]: I1227 07:49:15.331620 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://44124e136c1edec767fa7b55e82033376dbd4c05224d1c80d3b03b40d50a5a60" gracePeriod=600 Dec 27 07:49:15 crc kubenswrapper[4934]: I1227 07:49:15.847317 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="44124e136c1edec767fa7b55e82033376dbd4c05224d1c80d3b03b40d50a5a60" exitCode=0 Dec 27 07:49:15 crc kubenswrapper[4934]: I1227 07:49:15.847403 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"44124e136c1edec767fa7b55e82033376dbd4c05224d1c80d3b03b40d50a5a60"} Dec 27 07:49:15 crc kubenswrapper[4934]: I1227 07:49:15.847718 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"d237f7eb348f6dbacefb4e9dfc24791b0b82ed817b960e703b17f6f47221575e"} Dec 27 07:49:15 crc kubenswrapper[4934]: I1227 07:49:15.847742 4934 scope.go:117] "RemoveContainer" containerID="0219c2856889ae5331602db16e6c710418aca391266c15789503fdf6a068cf8c" Dec 27 07:49:17 crc kubenswrapper[4934]: I1227 07:49:17.468327 4934 scope.go:117] "RemoveContainer" containerID="fd35b2027f2556f022c6c16a393209f56220481bc406e1e57f5f9ac3f4134dfe" Dec 27 07:49:17 crc kubenswrapper[4934]: E1227 07:49:17.469484 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=prometheus pod=prometheus-k8s-0_openshift-monitoring(40c7cd23-14a7-4b30-9639-84d83ba74fe3)\"" pod="openshift-monitoring/prometheus-k8s-0" podUID="40c7cd23-14a7-4b30-9639-84d83ba74fe3" Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.324931 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" podUID="df922398-eb58-4f23-bfab-cfe9b90908c6" containerName="registry" containerID="cri-o://285a826b409cb26ca54b0f129254e7718f41a5ebe25c94a25c1ced5ddebe509d" gracePeriod=30 Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.850552 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.876640 4934 generic.go:334] "Generic (PLEG): container finished" podID="df922398-eb58-4f23-bfab-cfe9b90908c6" containerID="285a826b409cb26ca54b0f129254e7718f41a5ebe25c94a25c1ced5ddebe509d" exitCode=0 Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.876676 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" event={"ID":"df922398-eb58-4f23-bfab-cfe9b90908c6","Type":"ContainerDied","Data":"285a826b409cb26ca54b0f129254e7718f41a5ebe25c94a25c1ced5ddebe509d"} Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.876695 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" event={"ID":"df922398-eb58-4f23-bfab-cfe9b90908c6","Type":"ContainerDied","Data":"8f3350f6706e6a065226c875d01bf4bb025274fd53256fadead747481e7e1321"} Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.876713 4934 scope.go:117] "RemoveContainer" containerID="285a826b409cb26ca54b0f129254e7718f41a5ebe25c94a25c1ced5ddebe509d" Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.876809 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-ln99b" Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.894897 4934 scope.go:117] "RemoveContainer" containerID="285a826b409cb26ca54b0f129254e7718f41a5ebe25c94a25c1ced5ddebe509d" Dec 27 07:49:18 crc kubenswrapper[4934]: E1227 07:49:18.895453 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"285a826b409cb26ca54b0f129254e7718f41a5ebe25c94a25c1ced5ddebe509d\": container with ID starting with 285a826b409cb26ca54b0f129254e7718f41a5ebe25c94a25c1ced5ddebe509d not found: ID does not exist" containerID="285a826b409cb26ca54b0f129254e7718f41a5ebe25c94a25c1ced5ddebe509d" Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.895506 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"285a826b409cb26ca54b0f129254e7718f41a5ebe25c94a25c1ced5ddebe509d"} err="failed to get container status \"285a826b409cb26ca54b0f129254e7718f41a5ebe25c94a25c1ced5ddebe509d\": rpc error: code = NotFound desc = could not find container \"285a826b409cb26ca54b0f129254e7718f41a5ebe25c94a25c1ced5ddebe509d\": container with ID starting with 285a826b409cb26ca54b0f129254e7718f41a5ebe25c94a25c1ced5ddebe509d not found: ID does not exist" Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.925857 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/df922398-eb58-4f23-bfab-cfe9b90908c6-installation-pull-secrets\") pod \"df922398-eb58-4f23-bfab-cfe9b90908c6\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.925912 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/df922398-eb58-4f23-bfab-cfe9b90908c6-ca-trust-extracted\") pod \"df922398-eb58-4f23-bfab-cfe9b90908c6\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.925963 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/df922398-eb58-4f23-bfab-cfe9b90908c6-bound-sa-token\") pod \"df922398-eb58-4f23-bfab-cfe9b90908c6\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.926004 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/df922398-eb58-4f23-bfab-cfe9b90908c6-registry-certificates\") pod \"df922398-eb58-4f23-bfab-cfe9b90908c6\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.926190 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"df922398-eb58-4f23-bfab-cfe9b90908c6\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.926226 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9q69n\" (UniqueName: \"kubernetes.io/projected/df922398-eb58-4f23-bfab-cfe9b90908c6-kube-api-access-9q69n\") pod \"df922398-eb58-4f23-bfab-cfe9b90908c6\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.926248 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/df922398-eb58-4f23-bfab-cfe9b90908c6-registry-tls\") pod \"df922398-eb58-4f23-bfab-cfe9b90908c6\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.926282 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/df922398-eb58-4f23-bfab-cfe9b90908c6-trusted-ca\") pod \"df922398-eb58-4f23-bfab-cfe9b90908c6\" (UID: \"df922398-eb58-4f23-bfab-cfe9b90908c6\") " Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.927137 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df922398-eb58-4f23-bfab-cfe9b90908c6-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "df922398-eb58-4f23-bfab-cfe9b90908c6" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.927181 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df922398-eb58-4f23-bfab-cfe9b90908c6-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "df922398-eb58-4f23-bfab-cfe9b90908c6" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.932078 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df922398-eb58-4f23-bfab-cfe9b90908c6-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "df922398-eb58-4f23-bfab-cfe9b90908c6" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.932378 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df922398-eb58-4f23-bfab-cfe9b90908c6-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "df922398-eb58-4f23-bfab-cfe9b90908c6" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.935450 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df922398-eb58-4f23-bfab-cfe9b90908c6-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "df922398-eb58-4f23-bfab-cfe9b90908c6" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.935957 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df922398-eb58-4f23-bfab-cfe9b90908c6-kube-api-access-9q69n" (OuterVolumeSpecName: "kube-api-access-9q69n") pod "df922398-eb58-4f23-bfab-cfe9b90908c6" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6"). InnerVolumeSpecName "kube-api-access-9q69n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.937009 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "df922398-eb58-4f23-bfab-cfe9b90908c6" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 27 07:49:18 crc kubenswrapper[4934]: I1227 07:49:18.942945 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df922398-eb58-4f23-bfab-cfe9b90908c6-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "df922398-eb58-4f23-bfab-cfe9b90908c6" (UID: "df922398-eb58-4f23-bfab-cfe9b90908c6"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:49:19 crc kubenswrapper[4934]: I1227 07:49:19.027491 4934 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/df922398-eb58-4f23-bfab-cfe9b90908c6-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 27 07:49:19 crc kubenswrapper[4934]: I1227 07:49:19.027535 4934 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/df922398-eb58-4f23-bfab-cfe9b90908c6-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 27 07:49:19 crc kubenswrapper[4934]: I1227 07:49:19.027551 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9q69n\" (UniqueName: \"kubernetes.io/projected/df922398-eb58-4f23-bfab-cfe9b90908c6-kube-api-access-9q69n\") on node \"crc\" DevicePath \"\"" Dec 27 07:49:19 crc kubenswrapper[4934]: I1227 07:49:19.027562 4934 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/df922398-eb58-4f23-bfab-cfe9b90908c6-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 27 07:49:19 crc kubenswrapper[4934]: I1227 07:49:19.027573 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/df922398-eb58-4f23-bfab-cfe9b90908c6-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:49:19 crc kubenswrapper[4934]: I1227 07:49:19.027586 4934 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/df922398-eb58-4f23-bfab-cfe9b90908c6-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 27 07:49:19 crc kubenswrapper[4934]: I1227 07:49:19.027598 4934 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/df922398-eb58-4f23-bfab-cfe9b90908c6-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 27 07:49:19 crc kubenswrapper[4934]: I1227 07:49:19.231341 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ln99b"] Dec 27 07:49:19 crc kubenswrapper[4934]: I1227 07:49:19.240493 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-ln99b"] Dec 27 07:49:19 crc kubenswrapper[4934]: I1227 07:49:19.479673 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df922398-eb58-4f23-bfab-cfe9b90908c6" path="/var/lib/kubelet/pods/df922398-eb58-4f23-bfab-cfe9b90908c6/volumes" Dec 27 07:49:29 crc kubenswrapper[4934]: I1227 07:49:29.479189 4934 scope.go:117] "RemoveContainer" containerID="fd35b2027f2556f022c6c16a393209f56220481bc406e1e57f5f9ac3f4134dfe" Dec 27 07:49:29 crc kubenswrapper[4934]: I1227 07:49:29.994740 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_40c7cd23-14a7-4b30-9639-84d83ba74fe3/prometheus/2.log" Dec 27 07:49:29 crc kubenswrapper[4934]: I1227 07:49:29.996923 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"40c7cd23-14a7-4b30-9639-84d83ba74fe3","Type":"ContainerStarted","Data":"6fdf5d6ef349cc1a2b309f07be0f56b3129cf5751c93e1bf3dd5a2d571b00994"} Dec 27 07:49:30 crc kubenswrapper[4934]: I1227 07:49:30.030674 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-k8s-0" podStartSLOduration=56.469959607 podStartE2EDuration="1m0.030656205s" podCreationTimestamp="2025-12-27 07:48:30 +0000 UTC" firstStartedPulling="2025-12-27 07:48:32.224427516 +0000 UTC m=+373.044868100" lastFinishedPulling="2025-12-27 07:48:35.785124104 +0000 UTC m=+376.605564698" observedRunningTime="2025-12-27 07:49:30.029005659 +0000 UTC m=+430.849446253" watchObservedRunningTime="2025-12-27 07:49:30.030656205 +0000 UTC m=+430.851096799" Dec 27 07:49:30 crc kubenswrapper[4934]: I1227 07:49:30.952501 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:49:30 crc kubenswrapper[4934]: I1227 07:49:30.952862 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:49:31 crc kubenswrapper[4934]: I1227 07:49:31.001690 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:49:31 crc kubenswrapper[4934]: I1227 07:49:31.049533 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.533798 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5d9c4bb489-5mh5q"] Dec 27 07:50:10 crc kubenswrapper[4934]: E1227 07:50:10.535183 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="091c02c2-ae22-44e8-bf6b-64b69d6d9d83" containerName="console" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.535220 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="091c02c2-ae22-44e8-bf6b-64b69d6d9d83" containerName="console" Dec 27 07:50:10 crc kubenswrapper[4934]: E1227 07:50:10.535274 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df922398-eb58-4f23-bfab-cfe9b90908c6" containerName="registry" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.535296 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="df922398-eb58-4f23-bfab-cfe9b90908c6" containerName="registry" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.535560 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="df922398-eb58-4f23-bfab-cfe9b90908c6" containerName="registry" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.535594 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="091c02c2-ae22-44e8-bf6b-64b69d6d9d83" containerName="console" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.537962 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.555902 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5d9c4bb489-5mh5q"] Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.640561 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-trusted-ca-bundle\") pod \"console-5d9c4bb489-5mh5q\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.640873 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfz5v\" (UniqueName: \"kubernetes.io/projected/04fdcc85-c402-43ed-b1bc-4a11307e1640-kube-api-access-nfz5v\") pod \"console-5d9c4bb489-5mh5q\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.640939 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-service-ca\") pod \"console-5d9c4bb489-5mh5q\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.640971 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/04fdcc85-c402-43ed-b1bc-4a11307e1640-console-oauth-config\") pod \"console-5d9c4bb489-5mh5q\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.641129 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-console-config\") pod \"console-5d9c4bb489-5mh5q\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.641996 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/04fdcc85-c402-43ed-b1bc-4a11307e1640-console-serving-cert\") pod \"console-5d9c4bb489-5mh5q\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.642147 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-oauth-serving-cert\") pod \"console-5d9c4bb489-5mh5q\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.743258 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/04fdcc85-c402-43ed-b1bc-4a11307e1640-console-serving-cert\") pod \"console-5d9c4bb489-5mh5q\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.743302 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-oauth-serving-cert\") pod \"console-5d9c4bb489-5mh5q\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.743337 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-trusted-ca-bundle\") pod \"console-5d9c4bb489-5mh5q\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.743356 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfz5v\" (UniqueName: \"kubernetes.io/projected/04fdcc85-c402-43ed-b1bc-4a11307e1640-kube-api-access-nfz5v\") pod \"console-5d9c4bb489-5mh5q\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.744272 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-service-ca\") pod \"console-5d9c4bb489-5mh5q\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.744376 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/04fdcc85-c402-43ed-b1bc-4a11307e1640-console-oauth-config\") pod \"console-5d9c4bb489-5mh5q\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.744438 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-oauth-serving-cert\") pod \"console-5d9c4bb489-5mh5q\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.744519 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-console-config\") pod \"console-5d9c4bb489-5mh5q\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.745101 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-service-ca\") pod \"console-5d9c4bb489-5mh5q\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.745956 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-console-config\") pod \"console-5d9c4bb489-5mh5q\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.746263 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-trusted-ca-bundle\") pod \"console-5d9c4bb489-5mh5q\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.750879 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/04fdcc85-c402-43ed-b1bc-4a11307e1640-console-oauth-config\") pod \"console-5d9c4bb489-5mh5q\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.750905 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/04fdcc85-c402-43ed-b1bc-4a11307e1640-console-serving-cert\") pod \"console-5d9c4bb489-5mh5q\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.761551 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfz5v\" (UniqueName: \"kubernetes.io/projected/04fdcc85-c402-43ed-b1bc-4a11307e1640-kube-api-access-nfz5v\") pod \"console-5d9c4bb489-5mh5q\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:10 crc kubenswrapper[4934]: I1227 07:50:10.865742 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:11 crc kubenswrapper[4934]: I1227 07:50:11.300800 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5d9c4bb489-5mh5q"] Dec 27 07:50:12 crc kubenswrapper[4934]: I1227 07:50:12.310189 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5d9c4bb489-5mh5q" event={"ID":"04fdcc85-c402-43ed-b1bc-4a11307e1640","Type":"ContainerStarted","Data":"959bc9a219aee379abea83916b003f39b7a449d9588f9c1cbf8edaf25bb99a38"} Dec 27 07:50:12 crc kubenswrapper[4934]: I1227 07:50:12.310552 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5d9c4bb489-5mh5q" event={"ID":"04fdcc85-c402-43ed-b1bc-4a11307e1640","Type":"ContainerStarted","Data":"116de63b860bbe6f2887b3cf53331715a50d30e24f9ad06ec03115ba449a5d21"} Dec 27 07:50:12 crc kubenswrapper[4934]: I1227 07:50:12.336388 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5d9c4bb489-5mh5q" podStartSLOduration=2.336367612 podStartE2EDuration="2.336367612s" podCreationTimestamp="2025-12-27 07:50:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:50:12.33434046 +0000 UTC m=+473.154781084" watchObservedRunningTime="2025-12-27 07:50:12.336367612 +0000 UTC m=+473.156808246" Dec 27 07:50:20 crc kubenswrapper[4934]: I1227 07:50:20.866165 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:20 crc kubenswrapper[4934]: I1227 07:50:20.866828 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:20 crc kubenswrapper[4934]: I1227 07:50:20.874387 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:21 crc kubenswrapper[4934]: I1227 07:50:21.387565 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:50:21 crc kubenswrapper[4934]: I1227 07:50:21.479465 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-c8c89d49c-b97s8"] Dec 27 07:50:46 crc kubenswrapper[4934]: I1227 07:50:46.533030 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-c8c89d49c-b97s8" podUID="6aa1cec9-a0f2-46da-ae87-b94936a1360b" containerName="console" containerID="cri-o://b581d03cb9679949a3f4c7b52b451ceda693e476513699728786d21b2e99a932" gracePeriod=15 Dec 27 07:50:46 crc kubenswrapper[4934]: I1227 07:50:46.967323 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-c8c89d49c-b97s8_6aa1cec9-a0f2-46da-ae87-b94936a1360b/console/0.log" Dec 27 07:50:46 crc kubenswrapper[4934]: I1227 07:50:46.967706 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.164803 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2f5h\" (UniqueName: \"kubernetes.io/projected/6aa1cec9-a0f2-46da-ae87-b94936a1360b-kube-api-access-d2f5h\") pod \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.164904 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-oauth-serving-cert\") pod \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.164975 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-service-ca\") pod \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.165038 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-console-config\") pod \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.165286 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6aa1cec9-a0f2-46da-ae87-b94936a1360b-console-serving-cert\") pod \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.165411 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6aa1cec9-a0f2-46da-ae87-b94936a1360b-console-oauth-config\") pod \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.165483 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-trusted-ca-bundle\") pod \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\" (UID: \"6aa1cec9-a0f2-46da-ae87-b94936a1360b\") " Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.166457 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "6aa1cec9-a0f2-46da-ae87-b94936a1360b" (UID: "6aa1cec9-a0f2-46da-ae87-b94936a1360b"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.166482 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-service-ca" (OuterVolumeSpecName: "service-ca") pod "6aa1cec9-a0f2-46da-ae87-b94936a1360b" (UID: "6aa1cec9-a0f2-46da-ae87-b94936a1360b"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.167187 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6aa1cec9-a0f2-46da-ae87-b94936a1360b" (UID: "6aa1cec9-a0f2-46da-ae87-b94936a1360b"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.167662 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-console-config" (OuterVolumeSpecName: "console-config") pod "6aa1cec9-a0f2-46da-ae87-b94936a1360b" (UID: "6aa1cec9-a0f2-46da-ae87-b94936a1360b"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.170013 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aa1cec9-a0f2-46da-ae87-b94936a1360b-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "6aa1cec9-a0f2-46da-ae87-b94936a1360b" (UID: "6aa1cec9-a0f2-46da-ae87-b94936a1360b"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.170168 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aa1cec9-a0f2-46da-ae87-b94936a1360b-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "6aa1cec9-a0f2-46da-ae87-b94936a1360b" (UID: "6aa1cec9-a0f2-46da-ae87-b94936a1360b"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.170409 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6aa1cec9-a0f2-46da-ae87-b94936a1360b-kube-api-access-d2f5h" (OuterVolumeSpecName: "kube-api-access-d2f5h") pod "6aa1cec9-a0f2-46da-ae87-b94936a1360b" (UID: "6aa1cec9-a0f2-46da-ae87-b94936a1360b"). InnerVolumeSpecName "kube-api-access-d2f5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.268256 4934 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6aa1cec9-a0f2-46da-ae87-b94936a1360b-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.268318 4934 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6aa1cec9-a0f2-46da-ae87-b94936a1360b-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.268339 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.268363 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2f5h\" (UniqueName: \"kubernetes.io/projected/6aa1cec9-a0f2-46da-ae87-b94936a1360b-kube-api-access-d2f5h\") on node \"crc\" DevicePath \"\"" Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.268395 4934 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.268414 4934 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-service-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.268432 4934 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6aa1cec9-a0f2-46da-ae87-b94936a1360b-console-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.588551 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-c8c89d49c-b97s8_6aa1cec9-a0f2-46da-ae87-b94936a1360b/console/0.log" Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.588647 4934 generic.go:334] "Generic (PLEG): container finished" podID="6aa1cec9-a0f2-46da-ae87-b94936a1360b" containerID="b581d03cb9679949a3f4c7b52b451ceda693e476513699728786d21b2e99a932" exitCode=2 Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.588707 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-c8c89d49c-b97s8" event={"ID":"6aa1cec9-a0f2-46da-ae87-b94936a1360b","Type":"ContainerDied","Data":"b581d03cb9679949a3f4c7b52b451ceda693e476513699728786d21b2e99a932"} Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.588764 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-c8c89d49c-b97s8" event={"ID":"6aa1cec9-a0f2-46da-ae87-b94936a1360b","Type":"ContainerDied","Data":"a72702fbb69e3d117de98b84849a14e24c453d2440e5cdaf3901d3905dd77c92"} Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.588802 4934 scope.go:117] "RemoveContainer" containerID="b581d03cb9679949a3f4c7b52b451ceda693e476513699728786d21b2e99a932" Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.589024 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-c8c89d49c-b97s8" Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.620058 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-c8c89d49c-b97s8"] Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.628461 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-c8c89d49c-b97s8"] Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.631681 4934 scope.go:117] "RemoveContainer" containerID="b581d03cb9679949a3f4c7b52b451ceda693e476513699728786d21b2e99a932" Dec 27 07:50:47 crc kubenswrapper[4934]: E1227 07:50:47.632339 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b581d03cb9679949a3f4c7b52b451ceda693e476513699728786d21b2e99a932\": container with ID starting with b581d03cb9679949a3f4c7b52b451ceda693e476513699728786d21b2e99a932 not found: ID does not exist" containerID="b581d03cb9679949a3f4c7b52b451ceda693e476513699728786d21b2e99a932" Dec 27 07:50:47 crc kubenswrapper[4934]: I1227 07:50:47.632406 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b581d03cb9679949a3f4c7b52b451ceda693e476513699728786d21b2e99a932"} err="failed to get container status \"b581d03cb9679949a3f4c7b52b451ceda693e476513699728786d21b2e99a932\": rpc error: code = NotFound desc = could not find container \"b581d03cb9679949a3f4c7b52b451ceda693e476513699728786d21b2e99a932\": container with ID starting with b581d03cb9679949a3f4c7b52b451ceda693e476513699728786d21b2e99a932 not found: ID does not exist" Dec 27 07:50:49 crc kubenswrapper[4934]: I1227 07:50:49.483722 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6aa1cec9-a0f2-46da-ae87-b94936a1360b" path="/var/lib/kubelet/pods/6aa1cec9-a0f2-46da-ae87-b94936a1360b/volumes" Dec 27 07:51:15 crc kubenswrapper[4934]: I1227 07:51:15.330719 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 07:51:15 crc kubenswrapper[4934]: I1227 07:51:15.331252 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 07:51:45 crc kubenswrapper[4934]: I1227 07:51:45.330242 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 07:51:45 crc kubenswrapper[4934]: I1227 07:51:45.330731 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 07:52:15 crc kubenswrapper[4934]: I1227 07:52:15.330818 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 07:52:15 crc kubenswrapper[4934]: I1227 07:52:15.331536 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 07:52:15 crc kubenswrapper[4934]: I1227 07:52:15.331606 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 07:52:15 crc kubenswrapper[4934]: I1227 07:52:15.332582 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d237f7eb348f6dbacefb4e9dfc24791b0b82ed817b960e703b17f6f47221575e"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 07:52:15 crc kubenswrapper[4934]: I1227 07:52:15.332702 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://d237f7eb348f6dbacefb4e9dfc24791b0b82ed817b960e703b17f6f47221575e" gracePeriod=600 Dec 27 07:52:16 crc kubenswrapper[4934]: I1227 07:52:16.266009 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="d237f7eb348f6dbacefb4e9dfc24791b0b82ed817b960e703b17f6f47221575e" exitCode=0 Dec 27 07:52:16 crc kubenswrapper[4934]: I1227 07:52:16.266117 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"d237f7eb348f6dbacefb4e9dfc24791b0b82ed817b960e703b17f6f47221575e"} Dec 27 07:52:16 crc kubenswrapper[4934]: I1227 07:52:16.266680 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"56bf8d198d1688bbc71036fbf874230b44fa217f3818e5e1d2117e459b2b5477"} Dec 27 07:52:16 crc kubenswrapper[4934]: I1227 07:52:16.266710 4934 scope.go:117] "RemoveContainer" containerID="44124e136c1edec767fa7b55e82033376dbd4c05224d1c80d3b03b40d50a5a60" Dec 27 07:52:45 crc kubenswrapper[4934]: I1227 07:52:45.091753 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh"] Dec 27 07:52:45 crc kubenswrapper[4934]: E1227 07:52:45.092733 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aa1cec9-a0f2-46da-ae87-b94936a1360b" containerName="console" Dec 27 07:52:45 crc kubenswrapper[4934]: I1227 07:52:45.092757 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aa1cec9-a0f2-46da-ae87-b94936a1360b" containerName="console" Dec 27 07:52:45 crc kubenswrapper[4934]: I1227 07:52:45.092931 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aa1cec9-a0f2-46da-ae87-b94936a1360b" containerName="console" Dec 27 07:52:45 crc kubenswrapper[4934]: I1227 07:52:45.094243 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh" Dec 27 07:52:45 crc kubenswrapper[4934]: I1227 07:52:45.099341 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 27 07:52:45 crc kubenswrapper[4934]: I1227 07:52:45.104702 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh"] Dec 27 07:52:45 crc kubenswrapper[4934]: I1227 07:52:45.192733 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d1a4dc79-8563-47a8-afc8-fe018f9bb949-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh\" (UID: \"d1a4dc79-8563-47a8-afc8-fe018f9bb949\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh" Dec 27 07:52:45 crc kubenswrapper[4934]: I1227 07:52:45.192800 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t7z6\" (UniqueName: \"kubernetes.io/projected/d1a4dc79-8563-47a8-afc8-fe018f9bb949-kube-api-access-4t7z6\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh\" (UID: \"d1a4dc79-8563-47a8-afc8-fe018f9bb949\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh" Dec 27 07:52:45 crc kubenswrapper[4934]: I1227 07:52:45.192885 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d1a4dc79-8563-47a8-afc8-fe018f9bb949-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh\" (UID: \"d1a4dc79-8563-47a8-afc8-fe018f9bb949\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh" Dec 27 07:52:45 crc kubenswrapper[4934]: I1227 07:52:45.294420 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d1a4dc79-8563-47a8-afc8-fe018f9bb949-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh\" (UID: \"d1a4dc79-8563-47a8-afc8-fe018f9bb949\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh" Dec 27 07:52:45 crc kubenswrapper[4934]: I1227 07:52:45.294498 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d1a4dc79-8563-47a8-afc8-fe018f9bb949-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh\" (UID: \"d1a4dc79-8563-47a8-afc8-fe018f9bb949\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh" Dec 27 07:52:45 crc kubenswrapper[4934]: I1227 07:52:45.294543 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t7z6\" (UniqueName: \"kubernetes.io/projected/d1a4dc79-8563-47a8-afc8-fe018f9bb949-kube-api-access-4t7z6\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh\" (UID: \"d1a4dc79-8563-47a8-afc8-fe018f9bb949\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh" Dec 27 07:52:45 crc kubenswrapper[4934]: I1227 07:52:45.295144 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d1a4dc79-8563-47a8-afc8-fe018f9bb949-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh\" (UID: \"d1a4dc79-8563-47a8-afc8-fe018f9bb949\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh" Dec 27 07:52:45 crc kubenswrapper[4934]: I1227 07:52:45.295299 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d1a4dc79-8563-47a8-afc8-fe018f9bb949-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh\" (UID: \"d1a4dc79-8563-47a8-afc8-fe018f9bb949\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh" Dec 27 07:52:45 crc kubenswrapper[4934]: I1227 07:52:45.321125 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t7z6\" (UniqueName: \"kubernetes.io/projected/d1a4dc79-8563-47a8-afc8-fe018f9bb949-kube-api-access-4t7z6\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh\" (UID: \"d1a4dc79-8563-47a8-afc8-fe018f9bb949\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh" Dec 27 07:52:45 crc kubenswrapper[4934]: I1227 07:52:45.416906 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh" Dec 27 07:52:45 crc kubenswrapper[4934]: I1227 07:52:45.893154 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh"] Dec 27 07:52:46 crc kubenswrapper[4934]: I1227 07:52:46.507466 4934 generic.go:334] "Generic (PLEG): container finished" podID="d1a4dc79-8563-47a8-afc8-fe018f9bb949" containerID="c31d6409be35e904efba8eb167894ed5151b78736ed9a8b6747471e9fce61a75" exitCode=0 Dec 27 07:52:46 crc kubenswrapper[4934]: I1227 07:52:46.507509 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh" event={"ID":"d1a4dc79-8563-47a8-afc8-fe018f9bb949","Type":"ContainerDied","Data":"c31d6409be35e904efba8eb167894ed5151b78736ed9a8b6747471e9fce61a75"} Dec 27 07:52:46 crc kubenswrapper[4934]: I1227 07:52:46.507794 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh" event={"ID":"d1a4dc79-8563-47a8-afc8-fe018f9bb949","Type":"ContainerStarted","Data":"5dc18ec0e71f17eee6f8836f3595879ed4c0227382d04dc7fb7c357e7e7ccc49"} Dec 27 07:52:46 crc kubenswrapper[4934]: I1227 07:52:46.509757 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 27 07:52:48 crc kubenswrapper[4934]: I1227 07:52:48.523616 4934 generic.go:334] "Generic (PLEG): container finished" podID="d1a4dc79-8563-47a8-afc8-fe018f9bb949" containerID="dc43a5a479b29b6fa87ef5f45054043f650886c0036ec92defefccbe5a25313d" exitCode=0 Dec 27 07:52:48 crc kubenswrapper[4934]: I1227 07:52:48.523690 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh" event={"ID":"d1a4dc79-8563-47a8-afc8-fe018f9bb949","Type":"ContainerDied","Data":"dc43a5a479b29b6fa87ef5f45054043f650886c0036ec92defefccbe5a25313d"} Dec 27 07:52:49 crc kubenswrapper[4934]: I1227 07:52:49.532501 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh" event={"ID":"d1a4dc79-8563-47a8-afc8-fe018f9bb949","Type":"ContainerStarted","Data":"6e7ca89ab1dbe3a10062356ae3a34fbd9e3e69fe786c7db945942bf57469d28f"} Dec 27 07:52:49 crc kubenswrapper[4934]: I1227 07:52:49.554448 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh" podStartSLOduration=3.22755579 podStartE2EDuration="4.554427503s" podCreationTimestamp="2025-12-27 07:52:45 +0000 UTC" firstStartedPulling="2025-12-27 07:52:46.509543975 +0000 UTC m=+627.329984569" lastFinishedPulling="2025-12-27 07:52:47.836415688 +0000 UTC m=+628.656856282" observedRunningTime="2025-12-27 07:52:49.551510798 +0000 UTC m=+630.371951402" watchObservedRunningTime="2025-12-27 07:52:49.554427503 +0000 UTC m=+630.374868107" Dec 27 07:52:50 crc kubenswrapper[4934]: I1227 07:52:50.541441 4934 generic.go:334] "Generic (PLEG): container finished" podID="d1a4dc79-8563-47a8-afc8-fe018f9bb949" containerID="6e7ca89ab1dbe3a10062356ae3a34fbd9e3e69fe786c7db945942bf57469d28f" exitCode=0 Dec 27 07:52:50 crc kubenswrapper[4934]: I1227 07:52:50.541498 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh" event={"ID":"d1a4dc79-8563-47a8-afc8-fe018f9bb949","Type":"ContainerDied","Data":"6e7ca89ab1dbe3a10062356ae3a34fbd9e3e69fe786c7db945942bf57469d28f"} Dec 27 07:52:51 crc kubenswrapper[4934]: I1227 07:52:51.809308 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh" Dec 27 07:52:51 crc kubenswrapper[4934]: I1227 07:52:51.901133 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d1a4dc79-8563-47a8-afc8-fe018f9bb949-util\") pod \"d1a4dc79-8563-47a8-afc8-fe018f9bb949\" (UID: \"d1a4dc79-8563-47a8-afc8-fe018f9bb949\") " Dec 27 07:52:51 crc kubenswrapper[4934]: I1227 07:52:51.901241 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d1a4dc79-8563-47a8-afc8-fe018f9bb949-bundle\") pod \"d1a4dc79-8563-47a8-afc8-fe018f9bb949\" (UID: \"d1a4dc79-8563-47a8-afc8-fe018f9bb949\") " Dec 27 07:52:51 crc kubenswrapper[4934]: I1227 07:52:51.901304 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4t7z6\" (UniqueName: \"kubernetes.io/projected/d1a4dc79-8563-47a8-afc8-fe018f9bb949-kube-api-access-4t7z6\") pod \"d1a4dc79-8563-47a8-afc8-fe018f9bb949\" (UID: \"d1a4dc79-8563-47a8-afc8-fe018f9bb949\") " Dec 27 07:52:51 crc kubenswrapper[4934]: I1227 07:52:51.905721 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1a4dc79-8563-47a8-afc8-fe018f9bb949-bundle" (OuterVolumeSpecName: "bundle") pod "d1a4dc79-8563-47a8-afc8-fe018f9bb949" (UID: "d1a4dc79-8563-47a8-afc8-fe018f9bb949"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:52:51 crc kubenswrapper[4934]: I1227 07:52:51.907278 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1a4dc79-8563-47a8-afc8-fe018f9bb949-kube-api-access-4t7z6" (OuterVolumeSpecName: "kube-api-access-4t7z6") pod "d1a4dc79-8563-47a8-afc8-fe018f9bb949" (UID: "d1a4dc79-8563-47a8-afc8-fe018f9bb949"). InnerVolumeSpecName "kube-api-access-4t7z6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:52:52 crc kubenswrapper[4934]: I1227 07:52:52.003392 4934 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d1a4dc79-8563-47a8-afc8-fe018f9bb949-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:52 crc kubenswrapper[4934]: I1227 07:52:52.003423 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4t7z6\" (UniqueName: \"kubernetes.io/projected/d1a4dc79-8563-47a8-afc8-fe018f9bb949-kube-api-access-4t7z6\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:52 crc kubenswrapper[4934]: I1227 07:52:52.271888 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1a4dc79-8563-47a8-afc8-fe018f9bb949-util" (OuterVolumeSpecName: "util") pod "d1a4dc79-8563-47a8-afc8-fe018f9bb949" (UID: "d1a4dc79-8563-47a8-afc8-fe018f9bb949"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:52:52 crc kubenswrapper[4934]: I1227 07:52:52.309339 4934 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d1a4dc79-8563-47a8-afc8-fe018f9bb949-util\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:52 crc kubenswrapper[4934]: I1227 07:52:52.557542 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh" event={"ID":"d1a4dc79-8563-47a8-afc8-fe018f9bb949","Type":"ContainerDied","Data":"5dc18ec0e71f17eee6f8836f3595879ed4c0227382d04dc7fb7c357e7e7ccc49"} Dec 27 07:52:52 crc kubenswrapper[4934]: I1227 07:52:52.557890 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5dc18ec0e71f17eee6f8836f3595879ed4c0227382d04dc7fb7c357e7e7ccc49" Dec 27 07:52:52 crc kubenswrapper[4934]: I1227 07:52:52.557644 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh" Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.142927 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8vzdl"] Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.143553 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovn-controller" containerID="cri-o://b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6" gracePeriod=30 Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.143591 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18" gracePeriod=30 Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.143617 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="kube-rbac-proxy-node" containerID="cri-o://d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c" gracePeriod=30 Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.143671 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovn-acl-logging" containerID="cri-o://6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168" gracePeriod=30 Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.143692 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="northd" containerID="cri-o://857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5" gracePeriod=30 Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.143617 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="nbdb" containerID="cri-o://e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c" gracePeriod=30 Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.143805 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="sbdb" containerID="cri-o://d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f" gracePeriod=30 Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.186314 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovnkube-controller" containerID="cri-o://478607a19e4255f7af127be496f42f13353eac046569986baf8095c3860f53c9" gracePeriod=30 Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.587472 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovnkube-controller/3.log" Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.589891 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovn-acl-logging/0.log" Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.590488 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovn-controller/0.log" Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.590862 4934 generic.go:334] "Generic (PLEG): container finished" podID="1c4780a7-56b7-4772-af66-045a03b31a71" containerID="478607a19e4255f7af127be496f42f13353eac046569986baf8095c3860f53c9" exitCode=0 Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.590894 4934 generic.go:334] "Generic (PLEG): container finished" podID="1c4780a7-56b7-4772-af66-045a03b31a71" containerID="d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f" exitCode=0 Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.590903 4934 generic.go:334] "Generic (PLEG): container finished" podID="1c4780a7-56b7-4772-af66-045a03b31a71" containerID="e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c" exitCode=0 Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.590912 4934 generic.go:334] "Generic (PLEG): container finished" podID="1c4780a7-56b7-4772-af66-045a03b31a71" containerID="857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5" exitCode=0 Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.590924 4934 generic.go:334] "Generic (PLEG): container finished" podID="1c4780a7-56b7-4772-af66-045a03b31a71" containerID="6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168" exitCode=143 Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.590934 4934 generic.go:334] "Generic (PLEG): container finished" podID="1c4780a7-56b7-4772-af66-045a03b31a71" containerID="b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6" exitCode=143 Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.590936 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerDied","Data":"478607a19e4255f7af127be496f42f13353eac046569986baf8095c3860f53c9"} Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.590971 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerDied","Data":"d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f"} Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.590983 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerDied","Data":"e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c"} Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.590992 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerDied","Data":"857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5"} Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.591001 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerDied","Data":"6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168"} Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.591010 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerDied","Data":"b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6"} Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.591024 4934 scope.go:117] "RemoveContainer" containerID="0133bfa46253ea9cf0c58a1951b5b012f93f5e8e1ae0ce6bff2341e6d94366c4" Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.593073 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jzv2s_1bdd4b78-f615-4771-9448-8faa03dbc4f5/kube-multus/2.log" Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.593709 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jzv2s_1bdd4b78-f615-4771-9448-8faa03dbc4f5/kube-multus/1.log" Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.593757 4934 generic.go:334] "Generic (PLEG): container finished" podID="1bdd4b78-f615-4771-9448-8faa03dbc4f5" containerID="b3182ba625b9908dd1c4de2aa6395b330f542942567d562205e6ee11080b342a" exitCode=2 Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.593791 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jzv2s" event={"ID":"1bdd4b78-f615-4771-9448-8faa03dbc4f5","Type":"ContainerDied","Data":"b3182ba625b9908dd1c4de2aa6395b330f542942567d562205e6ee11080b342a"} Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.594383 4934 scope.go:117] "RemoveContainer" containerID="b3182ba625b9908dd1c4de2aa6395b330f542942567d562205e6ee11080b342a" Dec 27 07:52:56 crc kubenswrapper[4934]: E1227 07:52:56.594662 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-jzv2s_openshift-multus(1bdd4b78-f615-4771-9448-8faa03dbc4f5)\"" pod="openshift-multus/multus-jzv2s" podUID="1bdd4b78-f615-4771-9448-8faa03dbc4f5" Dec 27 07:52:56 crc kubenswrapper[4934]: I1227 07:52:56.635902 4934 scope.go:117] "RemoveContainer" containerID="dd93f8a550f1648117195c4caac3c1e150c84a1c2dba90ae1ba1a4ecc1bd284b" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.600044 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jzv2s_1bdd4b78-f615-4771-9448-8faa03dbc4f5/kube-multus/2.log" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.604389 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovn-acl-logging/0.log" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.604971 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovn-controller/0.log" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.605365 4934 generic.go:334] "Generic (PLEG): container finished" podID="1c4780a7-56b7-4772-af66-045a03b31a71" containerID="9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18" exitCode=0 Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.605387 4934 generic.go:334] "Generic (PLEG): container finished" podID="1c4780a7-56b7-4772-af66-045a03b31a71" containerID="d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c" exitCode=0 Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.605425 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerDied","Data":"9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18"} Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.605469 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerDied","Data":"d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c"} Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.832733 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovn-acl-logging/0.log" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.833566 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovn-controller/0.log" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.833975 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.891424 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qbcrd"] Dec 27 07:52:57 crc kubenswrapper[4934]: E1227 07:52:57.891651 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="sbdb" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.891663 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="sbdb" Dec 27 07:52:57 crc kubenswrapper[4934]: E1227 07:52:57.891672 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="kubecfg-setup" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.891678 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="kubecfg-setup" Dec 27 07:52:57 crc kubenswrapper[4934]: E1227 07:52:57.891688 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovnkube-controller" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.891694 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovnkube-controller" Dec 27 07:52:57 crc kubenswrapper[4934]: E1227 07:52:57.891702 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovnkube-controller" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.891708 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovnkube-controller" Dec 27 07:52:57 crc kubenswrapper[4934]: E1227 07:52:57.891717 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a4dc79-8563-47a8-afc8-fe018f9bb949" containerName="util" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.891723 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a4dc79-8563-47a8-afc8-fe018f9bb949" containerName="util" Dec 27 07:52:57 crc kubenswrapper[4934]: E1227 07:52:57.891731 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovn-acl-logging" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.891736 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovn-acl-logging" Dec 27 07:52:57 crc kubenswrapper[4934]: E1227 07:52:57.891743 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="nbdb" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.891749 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="nbdb" Dec 27 07:52:57 crc kubenswrapper[4934]: E1227 07:52:57.891757 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="northd" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.891763 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="northd" Dec 27 07:52:57 crc kubenswrapper[4934]: E1227 07:52:57.891776 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a4dc79-8563-47a8-afc8-fe018f9bb949" containerName="pull" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.891782 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a4dc79-8563-47a8-afc8-fe018f9bb949" containerName="pull" Dec 27 07:52:57 crc kubenswrapper[4934]: E1227 07:52:57.891788 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovnkube-controller" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.891793 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovnkube-controller" Dec 27 07:52:57 crc kubenswrapper[4934]: E1227 07:52:57.891800 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a4dc79-8563-47a8-afc8-fe018f9bb949" containerName="extract" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.891805 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a4dc79-8563-47a8-afc8-fe018f9bb949" containerName="extract" Dec 27 07:52:57 crc kubenswrapper[4934]: E1227 07:52:57.891814 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovnkube-controller" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.891819 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovnkube-controller" Dec 27 07:52:57 crc kubenswrapper[4934]: E1227 07:52:57.891828 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="kube-rbac-proxy-node" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.891834 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="kube-rbac-proxy-node" Dec 27 07:52:57 crc kubenswrapper[4934]: E1227 07:52:57.891839 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="kube-rbac-proxy-ovn-metrics" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.891844 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="kube-rbac-proxy-ovn-metrics" Dec 27 07:52:57 crc kubenswrapper[4934]: E1227 07:52:57.891856 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovn-controller" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.891862 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovn-controller" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.891959 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="northd" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.891967 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="nbdb" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.891977 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovn-controller" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.891984 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovnkube-controller" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.891992 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovnkube-controller" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.892000 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovn-acl-logging" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.892007 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovnkube-controller" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.892014 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1a4dc79-8563-47a8-afc8-fe018f9bb949" containerName="extract" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.892022 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="sbdb" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.892029 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="kube-rbac-proxy-ovn-metrics" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.892037 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="kube-rbac-proxy-node" Dec 27 07:52:57 crc kubenswrapper[4934]: E1227 07:52:57.892151 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovnkube-controller" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.892158 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovnkube-controller" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.892270 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovnkube-controller" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.892280 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" containerName="ovnkube-controller" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.893864 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.998816 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1c4780a7-56b7-4772-af66-045a03b31a71-ovn-node-metrics-cert\") pod \"1c4780a7-56b7-4772-af66-045a03b31a71\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.999049 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-kubelet\") pod \"1c4780a7-56b7-4772-af66-045a03b31a71\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.999242 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-cni-bin\") pod \"1c4780a7-56b7-4772-af66-045a03b31a71\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.999365 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-log-socket\") pod \"1c4780a7-56b7-4772-af66-045a03b31a71\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.999440 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1c4780a7-56b7-4772-af66-045a03b31a71-ovnkube-script-lib\") pod \"1c4780a7-56b7-4772-af66-045a03b31a71\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.999514 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-systemd-units\") pod \"1c4780a7-56b7-4772-af66-045a03b31a71\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.999570 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-cni-netd\") pod \"1c4780a7-56b7-4772-af66-045a03b31a71\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.999626 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-run-netns\") pod \"1c4780a7-56b7-4772-af66-045a03b31a71\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.999700 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-etc-openvswitch\") pod \"1c4780a7-56b7-4772-af66-045a03b31a71\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.999781 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-slash\") pod \"1c4780a7-56b7-4772-af66-045a03b31a71\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.999863 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-run-openvswitch\") pod \"1c4780a7-56b7-4772-af66-045a03b31a71\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " Dec 27 07:52:57 crc kubenswrapper[4934]: I1227 07:52:57.999928 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-var-lib-openvswitch\") pod \"1c4780a7-56b7-4772-af66-045a03b31a71\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:57.999999 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-run-ovn\") pod \"1c4780a7-56b7-4772-af66-045a03b31a71\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000071 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1c4780a7-56b7-4772-af66-045a03b31a71-ovnkube-config\") pod \"1c4780a7-56b7-4772-af66-045a03b31a71\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000156 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-node-log\") pod \"1c4780a7-56b7-4772-af66-045a03b31a71\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:57.999189 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "1c4780a7-56b7-4772-af66-045a03b31a71" (UID: "1c4780a7-56b7-4772-af66-045a03b31a71"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000198 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c4780a7-56b7-4772-af66-045a03b31a71-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "1c4780a7-56b7-4772-af66-045a03b31a71" (UID: "1c4780a7-56b7-4772-af66-045a03b31a71"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000271 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "1c4780a7-56b7-4772-af66-045a03b31a71" (UID: "1c4780a7-56b7-4772-af66-045a03b31a71"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000295 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-log-socket" (OuterVolumeSpecName: "log-socket") pod "1c4780a7-56b7-4772-af66-045a03b31a71" (UID: "1c4780a7-56b7-4772-af66-045a03b31a71"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000315 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-slash" (OuterVolumeSpecName: "host-slash") pod "1c4780a7-56b7-4772-af66-045a03b31a71" (UID: "1c4780a7-56b7-4772-af66-045a03b31a71"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000333 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "1c4780a7-56b7-4772-af66-045a03b31a71" (UID: "1c4780a7-56b7-4772-af66-045a03b31a71"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000352 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "1c4780a7-56b7-4772-af66-045a03b31a71" (UID: "1c4780a7-56b7-4772-af66-045a03b31a71"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000437 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kp9j\" (UniqueName: \"kubernetes.io/projected/1c4780a7-56b7-4772-af66-045a03b31a71-kube-api-access-2kp9j\") pod \"1c4780a7-56b7-4772-af66-045a03b31a71\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000504 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-var-lib-cni-networks-ovn-kubernetes\") pod \"1c4780a7-56b7-4772-af66-045a03b31a71\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000575 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-run-systemd\") pod \"1c4780a7-56b7-4772-af66-045a03b31a71\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000643 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-run-ovn-kubernetes\") pod \"1c4780a7-56b7-4772-af66-045a03b31a71\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000712 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1c4780a7-56b7-4772-af66-045a03b31a71-env-overrides\") pod \"1c4780a7-56b7-4772-af66-045a03b31a71\" (UID: \"1c4780a7-56b7-4772-af66-045a03b31a71\") " Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000896 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-host-cni-bin\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000978 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-host-run-netns\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.001052 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-run-openvswitch\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.001130 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-ovn-node-metrics-cert\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.001197 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-systemd-units\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.001270 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-host-slash\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.001335 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-host-cni-netd\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.001415 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-etc-openvswitch\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.001487 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-env-overrides\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.001550 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-node-log\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.001619 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-ovnkube-script-lib\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.001687 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-host-kubelet\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.001756 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-run-ovn\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.001822 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-host-run-ovn-kubernetes\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.001916 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-var-lib-openvswitch\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.001978 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-log-socket\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.002037 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-run-systemd\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.002123 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsd4t\" (UniqueName: \"kubernetes.io/projected/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-kube-api-access-fsd4t\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.002185 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-ovnkube-config\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.002264 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.002350 4934 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.002403 4934 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.002457 4934 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-log-socket\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.002524 4934 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1c4780a7-56b7-4772-af66-045a03b31a71-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.002585 4934 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.002641 4934 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.002692 4934 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-slash\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000458 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "1c4780a7-56b7-4772-af66-045a03b31a71" (UID: "1c4780a7-56b7-4772-af66-045a03b31a71"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000471 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "1c4780a7-56b7-4772-af66-045a03b31a71" (UID: "1c4780a7-56b7-4772-af66-045a03b31a71"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000490 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "1c4780a7-56b7-4772-af66-045a03b31a71" (UID: "1c4780a7-56b7-4772-af66-045a03b31a71"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000504 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "1c4780a7-56b7-4772-af66-045a03b31a71" (UID: "1c4780a7-56b7-4772-af66-045a03b31a71"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000516 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "1c4780a7-56b7-4772-af66-045a03b31a71" (UID: "1c4780a7-56b7-4772-af66-045a03b31a71"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000535 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-node-log" (OuterVolumeSpecName: "node-log") pod "1c4780a7-56b7-4772-af66-045a03b31a71" (UID: "1c4780a7-56b7-4772-af66-045a03b31a71"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000751 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c4780a7-56b7-4772-af66-045a03b31a71-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "1c4780a7-56b7-4772-af66-045a03b31a71" (UID: "1c4780a7-56b7-4772-af66-045a03b31a71"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.000773 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "1c4780a7-56b7-4772-af66-045a03b31a71" (UID: "1c4780a7-56b7-4772-af66-045a03b31a71"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.003199 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "1c4780a7-56b7-4772-af66-045a03b31a71" (UID: "1c4780a7-56b7-4772-af66-045a03b31a71"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.003570 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c4780a7-56b7-4772-af66-045a03b31a71-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "1c4780a7-56b7-4772-af66-045a03b31a71" (UID: "1c4780a7-56b7-4772-af66-045a03b31a71"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.004681 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c4780a7-56b7-4772-af66-045a03b31a71-kube-api-access-2kp9j" (OuterVolumeSpecName: "kube-api-access-2kp9j") pod "1c4780a7-56b7-4772-af66-045a03b31a71" (UID: "1c4780a7-56b7-4772-af66-045a03b31a71"). InnerVolumeSpecName "kube-api-access-2kp9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.004766 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c4780a7-56b7-4772-af66-045a03b31a71-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "1c4780a7-56b7-4772-af66-045a03b31a71" (UID: "1c4780a7-56b7-4772-af66-045a03b31a71"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.013598 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "1c4780a7-56b7-4772-af66-045a03b31a71" (UID: "1c4780a7-56b7-4772-af66-045a03b31a71"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.104681 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-var-lib-openvswitch\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.104726 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-log-socket\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.104747 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-run-systemd\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.104771 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsd4t\" (UniqueName: \"kubernetes.io/projected/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-kube-api-access-fsd4t\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.104781 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-var-lib-openvswitch\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.104792 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-ovnkube-config\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.104838 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.104867 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-log-socket\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.104903 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-host-cni-bin\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.104875 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-host-cni-bin\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.104938 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.104961 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-host-run-netns\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.104973 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-run-systemd\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105002 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-host-run-netns\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105013 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-run-openvswitch\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105032 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-ovn-node-metrics-cert\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105054 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-systemd-units\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105097 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-host-slash\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105115 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-host-cni-netd\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105168 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-etc-openvswitch\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105185 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-env-overrides\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105213 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-node-log\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105244 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-ovnkube-script-lib\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105270 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-host-kubelet\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105289 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-run-ovn\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105321 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-host-run-ovn-kubernetes\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105361 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-run-openvswitch\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105436 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-host-kubelet\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105438 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-run-ovn\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105456 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-host-run-ovn-kubernetes\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105464 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-host-cni-netd\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105474 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-etc-openvswitch\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105531 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-node-log\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105553 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-ovnkube-config\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105579 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-systemd-units\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105614 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-host-slash\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105658 4934 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1c4780a7-56b7-4772-af66-045a03b31a71-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105678 4934 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105691 4934 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105703 4934 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105717 4934 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105732 4934 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105743 4934 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1c4780a7-56b7-4772-af66-045a03b31a71-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105755 4934 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-node-log\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105766 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kp9j\" (UniqueName: \"kubernetes.io/projected/1c4780a7-56b7-4772-af66-045a03b31a71-kube-api-access-2kp9j\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105778 4934 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105791 4934 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105804 4934 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1c4780a7-56b7-4772-af66-045a03b31a71-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105805 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-env-overrides\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.105834 4934 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1c4780a7-56b7-4772-af66-045a03b31a71-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.106240 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-ovnkube-script-lib\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.108662 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-ovn-node-metrics-cert\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.121258 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsd4t\" (UniqueName: \"kubernetes.io/projected/9e6abdb0-68c6-4c92-b48d-e2df7520d18d-kube-api-access-fsd4t\") pod \"ovnkube-node-qbcrd\" (UID: \"9e6abdb0-68c6-4c92-b48d-e2df7520d18d\") " pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.205656 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.615910 4934 generic.go:334] "Generic (PLEG): container finished" podID="9e6abdb0-68c6-4c92-b48d-e2df7520d18d" containerID="6d1abb1c93300b52a8b1b54dca8f62cbdbe3a3a277eb60fac186b7440f6a6644" exitCode=0 Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.615985 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" event={"ID":"9e6abdb0-68c6-4c92-b48d-e2df7520d18d","Type":"ContainerDied","Data":"6d1abb1c93300b52a8b1b54dca8f62cbdbe3a3a277eb60fac186b7440f6a6644"} Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.616255 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" event={"ID":"9e6abdb0-68c6-4c92-b48d-e2df7520d18d","Type":"ContainerStarted","Data":"156fb8d9408c899635d27f6f3665ee3b8bbf18d9ccc014d33eb58cba18edf8a0"} Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.623527 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovn-acl-logging/0.log" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.624782 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8vzdl_1c4780a7-56b7-4772-af66-045a03b31a71/ovn-controller/0.log" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.625114 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" event={"ID":"1c4780a7-56b7-4772-af66-045a03b31a71","Type":"ContainerDied","Data":"18f078934d8adfcf20f5b70e00c58431e888b0eee3333dcf5da17bcf694a4684"} Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.625150 4934 scope.go:117] "RemoveContainer" containerID="478607a19e4255f7af127be496f42f13353eac046569986baf8095c3860f53c9" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.625286 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8vzdl" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.649916 4934 scope.go:117] "RemoveContainer" containerID="d07f1149982c2c8951a1e2afae99446631b385bad13325370656e50df3d4b67f" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.702655 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8vzdl"] Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.705344 4934 scope.go:117] "RemoveContainer" containerID="e0ed68ddd7e8deb0faf004558ac46c6159d2238cf5b777439579767c3a4f356c" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.713301 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8vzdl"] Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.734404 4934 scope.go:117] "RemoveContainer" containerID="857e61cf18424e2d8b8900578ed0fca1d30bdbfa27ef2880911035b7ada73cc5" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.759912 4934 scope.go:117] "RemoveContainer" containerID="9a71b9847a15ac2d174c795ce87b4fab6df97341e36cc6dbe94ed01a1e573a18" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.775191 4934 scope.go:117] "RemoveContainer" containerID="d00d0e7213ec47f046121f6782c0c4eac51eed3590534b1febd5fec4553f2e5c" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.802141 4934 scope.go:117] "RemoveContainer" containerID="6a7e759dbc186d2a329eb7f41562266a005bb78088157523adbc7293fdc21168" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.845818 4934 scope.go:117] "RemoveContainer" containerID="b2b2930a2962b84c3490b6ead60d709184521f634f5faa75377ba709ae49cdc6" Dec 27 07:52:58 crc kubenswrapper[4934]: I1227 07:52:58.875863 4934 scope.go:117] "RemoveContainer" containerID="92978ca10ba872ff2c368f669aed9d8466d3e9efb1a899db2fee96561d733bbd" Dec 27 07:52:59 crc kubenswrapper[4934]: I1227 07:52:59.473434 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c4780a7-56b7-4772-af66-045a03b31a71" path="/var/lib/kubelet/pods/1c4780a7-56b7-4772-af66-045a03b31a71/volumes" Dec 27 07:52:59 crc kubenswrapper[4934]: I1227 07:52:59.634203 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" event={"ID":"9e6abdb0-68c6-4c92-b48d-e2df7520d18d","Type":"ContainerStarted","Data":"7109a346770d2e79a3d6275459d62677d58664ab1ec6beb8c0515059d08b96f5"} Dec 27 07:52:59 crc kubenswrapper[4934]: I1227 07:52:59.634238 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" event={"ID":"9e6abdb0-68c6-4c92-b48d-e2df7520d18d","Type":"ContainerStarted","Data":"2cc230016f41dad502cd41a17769cdecd4f1e288b342f80571898be1bb51926a"} Dec 27 07:52:59 crc kubenswrapper[4934]: I1227 07:52:59.634248 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" event={"ID":"9e6abdb0-68c6-4c92-b48d-e2df7520d18d","Type":"ContainerStarted","Data":"156427747aa12ae84f2f667b08af59feff1cbfd58d1a9ef6c9356f1d9c53cbb9"} Dec 27 07:52:59 crc kubenswrapper[4934]: I1227 07:52:59.634256 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" event={"ID":"9e6abdb0-68c6-4c92-b48d-e2df7520d18d","Type":"ContainerStarted","Data":"ded2285ef5bc9aab8c9c3361e8c3a262ddb8196b0e1d0d7b134d1c3cfd23ddc8"} Dec 27 07:52:59 crc kubenswrapper[4934]: I1227 07:52:59.634265 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" event={"ID":"9e6abdb0-68c6-4c92-b48d-e2df7520d18d","Type":"ContainerStarted","Data":"51bbfc9bd71ccc135f3a6266f3e6e5006d0320286b0100cf716eac40e717dffb"} Dec 27 07:52:59 crc kubenswrapper[4934]: I1227 07:52:59.634274 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" event={"ID":"9e6abdb0-68c6-4c92-b48d-e2df7520d18d","Type":"ContainerStarted","Data":"c050bdc6cc1472714000063715b98a1954bb3d9c440d0ba41c6f475cb3a410b8"} Dec 27 07:53:02 crc kubenswrapper[4934]: I1227 07:53:02.655927 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" event={"ID":"9e6abdb0-68c6-4c92-b48d-e2df7520d18d","Type":"ContainerStarted","Data":"1f705631550b01c65034b877698fa6697ea2b296463d1e0223f91a50602948c0"} Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.002016 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6"] Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.003219 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.004925 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-gfgsx" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.006157 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.015134 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.081566 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grrff\" (UniqueName: \"kubernetes.io/projected/be52e823-2384-41da-b213-85e7fe61250e-kube-api-access-grrff\") pod \"obo-prometheus-operator-68bc856cb9-8llj6\" (UID: \"be52e823-2384-41da-b213-85e7fe61250e\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.134450 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn"] Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.135379 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.137546 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-kxp2f" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.147715 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.165289 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8"] Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.165985 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.182985 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grrff\" (UniqueName: \"kubernetes.io/projected/be52e823-2384-41da-b213-85e7fe61250e-kube-api-access-grrff\") pod \"obo-prometheus-operator-68bc856cb9-8llj6\" (UID: \"be52e823-2384-41da-b213-85e7fe61250e\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.201916 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grrff\" (UniqueName: \"kubernetes.io/projected/be52e823-2384-41da-b213-85e7fe61250e-kube-api-access-grrff\") pod \"obo-prometheus-operator-68bc856cb9-8llj6\" (UID: \"be52e823-2384-41da-b213-85e7fe61250e\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.284568 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/084dcdaf-73d5-45a7-9d98-a1dbcd76db41-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8\" (UID: \"084dcdaf-73d5-45a7-9d98-a1dbcd76db41\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.284630 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/084dcdaf-73d5-45a7-9d98-a1dbcd76db41-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8\" (UID: \"084dcdaf-73d5-45a7-9d98-a1dbcd76db41\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.284695 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b86f501c-2f1d-4131-820b-c9b71c4973dc-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn\" (UID: \"b86f501c-2f1d-4131-820b-c9b71c4973dc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.284747 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b86f501c-2f1d-4131-820b-c9b71c4973dc-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn\" (UID: \"b86f501c-2f1d-4131-820b-c9b71c4973dc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.318473 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.345712 4934 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-8llj6_openshift-operators_be52e823-2384-41da-b213-85e7fe61250e_0(58e5210a2029800fb0c64aa0d59ed01f5d7b71651deb1b5efa811b3481661b8d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.345787 4934 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-8llj6_openshift-operators_be52e823-2384-41da-b213-85e7fe61250e_0(58e5210a2029800fb0c64aa0d59ed01f5d7b71651deb1b5efa811b3481661b8d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.345810 4934 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-8llj6_openshift-operators_be52e823-2384-41da-b213-85e7fe61250e_0(58e5210a2029800fb0c64aa0d59ed01f5d7b71651deb1b5efa811b3481661b8d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.345867 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-68bc856cb9-8llj6_openshift-operators(be52e823-2384-41da-b213-85e7fe61250e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-68bc856cb9-8llj6_openshift-operators(be52e823-2384-41da-b213-85e7fe61250e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-8llj6_openshift-operators_be52e823-2384-41da-b213-85e7fe61250e_0(58e5210a2029800fb0c64aa0d59ed01f5d7b71651deb1b5efa811b3481661b8d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" podUID="be52e823-2384-41da-b213-85e7fe61250e" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.348217 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-lnknm"] Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.348943 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.350406 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.350540 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-m2bpv" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.386644 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/084dcdaf-73d5-45a7-9d98-a1dbcd76db41-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8\" (UID: \"084dcdaf-73d5-45a7-9d98-a1dbcd76db41\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.386712 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/084dcdaf-73d5-45a7-9d98-a1dbcd76db41-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8\" (UID: \"084dcdaf-73d5-45a7-9d98-a1dbcd76db41\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.386764 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b86f501c-2f1d-4131-820b-c9b71c4973dc-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn\" (UID: \"b86f501c-2f1d-4131-820b-c9b71c4973dc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.386826 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b86f501c-2f1d-4131-820b-c9b71c4973dc-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn\" (UID: \"b86f501c-2f1d-4131-820b-c9b71c4973dc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.390615 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b86f501c-2f1d-4131-820b-c9b71c4973dc-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn\" (UID: \"b86f501c-2f1d-4131-820b-c9b71c4973dc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.390907 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/084dcdaf-73d5-45a7-9d98-a1dbcd76db41-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8\" (UID: \"084dcdaf-73d5-45a7-9d98-a1dbcd76db41\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.391382 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/084dcdaf-73d5-45a7-9d98-a1dbcd76db41-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8\" (UID: \"084dcdaf-73d5-45a7-9d98-a1dbcd76db41\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.391689 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b86f501c-2f1d-4131-820b-c9b71c4973dc-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn\" (UID: \"b86f501c-2f1d-4131-820b-c9b71c4973dc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.451185 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.475318 4934 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn_openshift-operators_b86f501c-2f1d-4131-820b-c9b71c4973dc_0(fda3ec33184f9953255844330053e7331d3da654dfd3181ffebb6d7034d811a8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.475378 4934 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn_openshift-operators_b86f501c-2f1d-4131-820b-c9b71c4973dc_0(fda3ec33184f9953255844330053e7331d3da654dfd3181ffebb6d7034d811a8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.475400 4934 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn_openshift-operators_b86f501c-2f1d-4131-820b-c9b71c4973dc_0(fda3ec33184f9953255844330053e7331d3da654dfd3181ffebb6d7034d811a8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.475439 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn_openshift-operators(b86f501c-2f1d-4131-820b-c9b71c4973dc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn_openshift-operators(b86f501c-2f1d-4131-820b-c9b71c4973dc)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn_openshift-operators_b86f501c-2f1d-4131-820b-c9b71c4973dc_0(fda3ec33184f9953255844330053e7331d3da654dfd3181ffebb6d7034d811a8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" podUID="b86f501c-2f1d-4131-820b-c9b71c4973dc" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.479194 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.492265 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/4575d285-879f-4cc3-a9b9-406dc96e852d-observability-operator-tls\") pod \"observability-operator-59bdc8b94-lnknm\" (UID: \"4575d285-879f-4cc3-a9b9-406dc96e852d\") " pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.492373 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zchj\" (UniqueName: \"kubernetes.io/projected/4575d285-879f-4cc3-a9b9-406dc96e852d-kube-api-access-4zchj\") pod \"observability-operator-59bdc8b94-lnknm\" (UID: \"4575d285-879f-4cc3-a9b9-406dc96e852d\") " pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.502696 4934 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8_openshift-operators_084dcdaf-73d5-45a7-9d98-a1dbcd76db41_0(eb5741a218d75f0cc225cff36d739a39013e7fe37b24c3aaafd9f3192f0148bd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.502750 4934 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8_openshift-operators_084dcdaf-73d5-45a7-9d98-a1dbcd76db41_0(eb5741a218d75f0cc225cff36d739a39013e7fe37b24c3aaafd9f3192f0148bd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.502774 4934 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8_openshift-operators_084dcdaf-73d5-45a7-9d98-a1dbcd76db41_0(eb5741a218d75f0cc225cff36d739a39013e7fe37b24c3aaafd9f3192f0148bd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.502818 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8_openshift-operators(084dcdaf-73d5-45a7-9d98-a1dbcd76db41)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8_openshift-operators(084dcdaf-73d5-45a7-9d98-a1dbcd76db41)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8_openshift-operators_084dcdaf-73d5-45a7-9d98-a1dbcd76db41_0(eb5741a218d75f0cc225cff36d739a39013e7fe37b24c3aaafd9f3192f0148bd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" podUID="084dcdaf-73d5-45a7-9d98-a1dbcd76db41" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.518620 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-tx6z5"] Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.520240 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.524563 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-4g92z" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.593342 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wws5\" (UniqueName: \"kubernetes.io/projected/f97e8d86-1e9e-4c5a-a164-2600193c576b-kube-api-access-2wws5\") pod \"perses-operator-5bf474d74f-tx6z5\" (UID: \"f97e8d86-1e9e-4c5a-a164-2600193c576b\") " pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.593389 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/f97e8d86-1e9e-4c5a-a164-2600193c576b-openshift-service-ca\") pod \"perses-operator-5bf474d74f-tx6z5\" (UID: \"f97e8d86-1e9e-4c5a-a164-2600193c576b\") " pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.593421 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/4575d285-879f-4cc3-a9b9-406dc96e852d-observability-operator-tls\") pod \"observability-operator-59bdc8b94-lnknm\" (UID: \"4575d285-879f-4cc3-a9b9-406dc96e852d\") " pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.593528 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zchj\" (UniqueName: \"kubernetes.io/projected/4575d285-879f-4cc3-a9b9-406dc96e852d-kube-api-access-4zchj\") pod \"observability-operator-59bdc8b94-lnknm\" (UID: \"4575d285-879f-4cc3-a9b9-406dc96e852d\") " pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.597428 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/4575d285-879f-4cc3-a9b9-406dc96e852d-observability-operator-tls\") pod \"observability-operator-59bdc8b94-lnknm\" (UID: \"4575d285-879f-4cc3-a9b9-406dc96e852d\") " pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.633047 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zchj\" (UniqueName: \"kubernetes.io/projected/4575d285-879f-4cc3-a9b9-406dc96e852d-kube-api-access-4zchj\") pod \"observability-operator-59bdc8b94-lnknm\" (UID: \"4575d285-879f-4cc3-a9b9-406dc96e852d\") " pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.668223 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.673427 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" event={"ID":"9e6abdb0-68c6-4c92-b48d-e2df7520d18d","Type":"ContainerStarted","Data":"127595ed0c7738a3e25b54fc3d90fb63d7a038f5772f519f069e1df49472950e"} Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.673753 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.673891 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.674053 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.694383 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wws5\" (UniqueName: \"kubernetes.io/projected/f97e8d86-1e9e-4c5a-a164-2600193c576b-kube-api-access-2wws5\") pod \"perses-operator-5bf474d74f-tx6z5\" (UID: \"f97e8d86-1e9e-4c5a-a164-2600193c576b\") " pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.694751 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/f97e8d86-1e9e-4c5a-a164-2600193c576b-openshift-service-ca\") pod \"perses-operator-5bf474d74f-tx6z5\" (UID: \"f97e8d86-1e9e-4c5a-a164-2600193c576b\") " pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.695789 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/f97e8d86-1e9e-4c5a-a164-2600193c576b-openshift-service-ca\") pod \"perses-operator-5bf474d74f-tx6z5\" (UID: \"f97e8d86-1e9e-4c5a-a164-2600193c576b\") " pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.705169 4934 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-lnknm_openshift-operators_4575d285-879f-4cc3-a9b9-406dc96e852d_0(8d35f26ca707d7b1008268cf44268a065da079e506c9b27b435c68b3433090ad): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.705248 4934 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-lnknm_openshift-operators_4575d285-879f-4cc3-a9b9-406dc96e852d_0(8d35f26ca707d7b1008268cf44268a065da079e506c9b27b435c68b3433090ad): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.705275 4934 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-lnknm_openshift-operators_4575d285-879f-4cc3-a9b9-406dc96e852d_0(8d35f26ca707d7b1008268cf44268a065da079e506c9b27b435c68b3433090ad): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.705329 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-59bdc8b94-lnknm_openshift-operators(4575d285-879f-4cc3-a9b9-406dc96e852d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-59bdc8b94-lnknm_openshift-operators(4575d285-879f-4cc3-a9b9-406dc96e852d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-lnknm_openshift-operators_4575d285-879f-4cc3-a9b9-406dc96e852d_0(8d35f26ca707d7b1008268cf44268a065da079e506c9b27b435c68b3433090ad): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" podUID="4575d285-879f-4cc3-a9b9-406dc96e852d" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.708942 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.715875 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wws5\" (UniqueName: \"kubernetes.io/projected/f97e8d86-1e9e-4c5a-a164-2600193c576b-kube-api-access-2wws5\") pod \"perses-operator-5bf474d74f-tx6z5\" (UID: \"f97e8d86-1e9e-4c5a-a164-2600193c576b\") " pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.720382 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.721111 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" podStartSLOduration=7.721067095 podStartE2EDuration="7.721067095s" podCreationTimestamp="2025-12-27 07:52:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:53:04.720587412 +0000 UTC m=+645.541028006" watchObservedRunningTime="2025-12-27 07:53:04.721067095 +0000 UTC m=+645.541507689" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.835123 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.875260 4934 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-tx6z5_openshift-operators_f97e8d86-1e9e-4c5a-a164-2600193c576b_0(95c0a212624c452fa8394296c66a2d4e56b42daf402eeaf8eac35dd5f2d46a48): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.875328 4934 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-tx6z5_openshift-operators_f97e8d86-1e9e-4c5a-a164-2600193c576b_0(95c0a212624c452fa8394296c66a2d4e56b42daf402eeaf8eac35dd5f2d46a48): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.875351 4934 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-tx6z5_openshift-operators_f97e8d86-1e9e-4c5a-a164-2600193c576b_0(95c0a212624c452fa8394296c66a2d4e56b42daf402eeaf8eac35dd5f2d46a48): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.875395 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5bf474d74f-tx6z5_openshift-operators(f97e8d86-1e9e-4c5a-a164-2600193c576b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5bf474d74f-tx6z5_openshift-operators(f97e8d86-1e9e-4c5a-a164-2600193c576b)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-tx6z5_openshift-operators_f97e8d86-1e9e-4c5a-a164-2600193c576b_0(95c0a212624c452fa8394296c66a2d4e56b42daf402eeaf8eac35dd5f2d46a48): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" podUID="f97e8d86-1e9e-4c5a-a164-2600193c576b" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.939125 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8"] Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.939256 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.939708 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.942858 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6"] Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.942983 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.943438 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.952385 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-tx6z5"] Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.974797 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-lnknm"] Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.975267 4934 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-8llj6_openshift-operators_be52e823-2384-41da-b213-85e7fe61250e_0(668bb2ed65a4007b9ff6a914d57aff0eef535232b4901bc8073eb13cf468e834): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.975345 4934 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-8llj6_openshift-operators_be52e823-2384-41da-b213-85e7fe61250e_0(668bb2ed65a4007b9ff6a914d57aff0eef535232b4901bc8073eb13cf468e834): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.975369 4934 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-8llj6_openshift-operators_be52e823-2384-41da-b213-85e7fe61250e_0(668bb2ed65a4007b9ff6a914d57aff0eef535232b4901bc8073eb13cf468e834): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.975414 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-68bc856cb9-8llj6_openshift-operators(be52e823-2384-41da-b213-85e7fe61250e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-68bc856cb9-8llj6_openshift-operators(be52e823-2384-41da-b213-85e7fe61250e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-8llj6_openshift-operators_be52e823-2384-41da-b213-85e7fe61250e_0(668bb2ed65a4007b9ff6a914d57aff0eef535232b4901bc8073eb13cf468e834): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" podUID="be52e823-2384-41da-b213-85e7fe61250e" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.981682 4934 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8_openshift-operators_084dcdaf-73d5-45a7-9d98-a1dbcd76db41_0(3cfcb92ca14c6cb7a6005c36f5d0caaca029523775fb33a65ab4d8360648df9b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.981738 4934 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8_openshift-operators_084dcdaf-73d5-45a7-9d98-a1dbcd76db41_0(3cfcb92ca14c6cb7a6005c36f5d0caaca029523775fb33a65ab4d8360648df9b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.981757 4934 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8_openshift-operators_084dcdaf-73d5-45a7-9d98-a1dbcd76db41_0(3cfcb92ca14c6cb7a6005c36f5d0caaca029523775fb33a65ab4d8360648df9b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" Dec 27 07:53:04 crc kubenswrapper[4934]: E1227 07:53:04.981797 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8_openshift-operators(084dcdaf-73d5-45a7-9d98-a1dbcd76db41)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8_openshift-operators(084dcdaf-73d5-45a7-9d98-a1dbcd76db41)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8_openshift-operators_084dcdaf-73d5-45a7-9d98-a1dbcd76db41_0(3cfcb92ca14c6cb7a6005c36f5d0caaca029523775fb33a65ab4d8360648df9b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" podUID="084dcdaf-73d5-45a7-9d98-a1dbcd76db41" Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.999640 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn"] Dec 27 07:53:04 crc kubenswrapper[4934]: I1227 07:53:04.999740 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" Dec 27 07:53:05 crc kubenswrapper[4934]: I1227 07:53:05.000206 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" Dec 27 07:53:05 crc kubenswrapper[4934]: E1227 07:53:05.023254 4934 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn_openshift-operators_b86f501c-2f1d-4131-820b-c9b71c4973dc_0(dcf3beaedbc73e71d992109ea7dfb895c15fcbd2360f4d0c92337fb0dba8e28b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 27 07:53:05 crc kubenswrapper[4934]: E1227 07:53:05.023323 4934 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn_openshift-operators_b86f501c-2f1d-4131-820b-c9b71c4973dc_0(dcf3beaedbc73e71d992109ea7dfb895c15fcbd2360f4d0c92337fb0dba8e28b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" Dec 27 07:53:05 crc kubenswrapper[4934]: E1227 07:53:05.023343 4934 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn_openshift-operators_b86f501c-2f1d-4131-820b-c9b71c4973dc_0(dcf3beaedbc73e71d992109ea7dfb895c15fcbd2360f4d0c92337fb0dba8e28b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" Dec 27 07:53:05 crc kubenswrapper[4934]: E1227 07:53:05.023387 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn_openshift-operators(b86f501c-2f1d-4131-820b-c9b71c4973dc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn_openshift-operators(b86f501c-2f1d-4131-820b-c9b71c4973dc)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn_openshift-operators_b86f501c-2f1d-4131-820b-c9b71c4973dc_0(dcf3beaedbc73e71d992109ea7dfb895c15fcbd2360f4d0c92337fb0dba8e28b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" podUID="b86f501c-2f1d-4131-820b-c9b71c4973dc" Dec 27 07:53:05 crc kubenswrapper[4934]: I1227 07:53:05.680133 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:05 crc kubenswrapper[4934]: I1227 07:53:05.680184 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:05 crc kubenswrapper[4934]: I1227 07:53:05.680851 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:05 crc kubenswrapper[4934]: I1227 07:53:05.681188 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:05 crc kubenswrapper[4934]: E1227 07:53:05.721914 4934 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-tx6z5_openshift-operators_f97e8d86-1e9e-4c5a-a164-2600193c576b_0(0f5b47fbbd3f38a30c041ea49a73c462c9f648c66f8f296c7bc10c371be54f47): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 27 07:53:05 crc kubenswrapper[4934]: E1227 07:53:05.721975 4934 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-tx6z5_openshift-operators_f97e8d86-1e9e-4c5a-a164-2600193c576b_0(0f5b47fbbd3f38a30c041ea49a73c462c9f648c66f8f296c7bc10c371be54f47): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:05 crc kubenswrapper[4934]: E1227 07:53:05.721998 4934 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-tx6z5_openshift-operators_f97e8d86-1e9e-4c5a-a164-2600193c576b_0(0f5b47fbbd3f38a30c041ea49a73c462c9f648c66f8f296c7bc10c371be54f47): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:05 crc kubenswrapper[4934]: E1227 07:53:05.722042 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5bf474d74f-tx6z5_openshift-operators(f97e8d86-1e9e-4c5a-a164-2600193c576b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5bf474d74f-tx6z5_openshift-operators(f97e8d86-1e9e-4c5a-a164-2600193c576b)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-tx6z5_openshift-operators_f97e8d86-1e9e-4c5a-a164-2600193c576b_0(0f5b47fbbd3f38a30c041ea49a73c462c9f648c66f8f296c7bc10c371be54f47): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" podUID="f97e8d86-1e9e-4c5a-a164-2600193c576b" Dec 27 07:53:05 crc kubenswrapper[4934]: E1227 07:53:05.727356 4934 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-lnknm_openshift-operators_4575d285-879f-4cc3-a9b9-406dc96e852d_0(d3adfdd3ed6f1169d1d1c2e062810c8101b1cc33ab3adf3fc962a223d4c0484a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 27 07:53:05 crc kubenswrapper[4934]: E1227 07:53:05.727423 4934 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-lnknm_openshift-operators_4575d285-879f-4cc3-a9b9-406dc96e852d_0(d3adfdd3ed6f1169d1d1c2e062810c8101b1cc33ab3adf3fc962a223d4c0484a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:05 crc kubenswrapper[4934]: E1227 07:53:05.727446 4934 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-lnknm_openshift-operators_4575d285-879f-4cc3-a9b9-406dc96e852d_0(d3adfdd3ed6f1169d1d1c2e062810c8101b1cc33ab3adf3fc962a223d4c0484a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:05 crc kubenswrapper[4934]: E1227 07:53:05.727489 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-59bdc8b94-lnknm_openshift-operators(4575d285-879f-4cc3-a9b9-406dc96e852d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-59bdc8b94-lnknm_openshift-operators(4575d285-879f-4cc3-a9b9-406dc96e852d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-lnknm_openshift-operators_4575d285-879f-4cc3-a9b9-406dc96e852d_0(d3adfdd3ed6f1169d1d1c2e062810c8101b1cc33ab3adf3fc962a223d4c0484a): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" podUID="4575d285-879f-4cc3-a9b9-406dc96e852d" Dec 27 07:53:08 crc kubenswrapper[4934]: I1227 07:53:08.468434 4934 scope.go:117] "RemoveContainer" containerID="b3182ba625b9908dd1c4de2aa6395b330f542942567d562205e6ee11080b342a" Dec 27 07:53:08 crc kubenswrapper[4934]: E1227 07:53:08.469076 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-jzv2s_openshift-multus(1bdd4b78-f615-4771-9448-8faa03dbc4f5)\"" pod="openshift-multus/multus-jzv2s" podUID="1bdd4b78-f615-4771-9448-8faa03dbc4f5" Dec 27 07:53:15 crc kubenswrapper[4934]: I1227 07:53:15.467363 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" Dec 27 07:53:15 crc kubenswrapper[4934]: I1227 07:53:15.468541 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" Dec 27 07:53:15 crc kubenswrapper[4934]: E1227 07:53:15.523671 4934 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn_openshift-operators_b86f501c-2f1d-4131-820b-c9b71c4973dc_0(2eb989c29b624098b1fa9feee561461a02ca852133a5d107ef35ff8cc93bbffb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 27 07:53:15 crc kubenswrapper[4934]: E1227 07:53:15.524044 4934 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn_openshift-operators_b86f501c-2f1d-4131-820b-c9b71c4973dc_0(2eb989c29b624098b1fa9feee561461a02ca852133a5d107ef35ff8cc93bbffb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" Dec 27 07:53:15 crc kubenswrapper[4934]: E1227 07:53:15.524186 4934 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn_openshift-operators_b86f501c-2f1d-4131-820b-c9b71c4973dc_0(2eb989c29b624098b1fa9feee561461a02ca852133a5d107ef35ff8cc93bbffb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" Dec 27 07:53:15 crc kubenswrapper[4934]: E1227 07:53:15.524275 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn_openshift-operators(b86f501c-2f1d-4131-820b-c9b71c4973dc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn_openshift-operators(b86f501c-2f1d-4131-820b-c9b71c4973dc)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn_openshift-operators_b86f501c-2f1d-4131-820b-c9b71c4973dc_0(2eb989c29b624098b1fa9feee561461a02ca852133a5d107ef35ff8cc93bbffb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" podUID="b86f501c-2f1d-4131-820b-c9b71c4973dc" Dec 27 07:53:17 crc kubenswrapper[4934]: I1227 07:53:17.467031 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" Dec 27 07:53:17 crc kubenswrapper[4934]: I1227 07:53:17.467495 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" Dec 27 07:53:17 crc kubenswrapper[4934]: E1227 07:53:17.492808 4934 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8_openshift-operators_084dcdaf-73d5-45a7-9d98-a1dbcd76db41_0(31fa5097e539b821bba0a096384834e14188605d9c87e1c2a45e4bae8c1f07a7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 27 07:53:17 crc kubenswrapper[4934]: E1227 07:53:17.493228 4934 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8_openshift-operators_084dcdaf-73d5-45a7-9d98-a1dbcd76db41_0(31fa5097e539b821bba0a096384834e14188605d9c87e1c2a45e4bae8c1f07a7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" Dec 27 07:53:17 crc kubenswrapper[4934]: E1227 07:53:17.493254 4934 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8_openshift-operators_084dcdaf-73d5-45a7-9d98-a1dbcd76db41_0(31fa5097e539b821bba0a096384834e14188605d9c87e1c2a45e4bae8c1f07a7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" Dec 27 07:53:17 crc kubenswrapper[4934]: E1227 07:53:17.493312 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8_openshift-operators(084dcdaf-73d5-45a7-9d98-a1dbcd76db41)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8_openshift-operators(084dcdaf-73d5-45a7-9d98-a1dbcd76db41)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8_openshift-operators_084dcdaf-73d5-45a7-9d98-a1dbcd76db41_0(31fa5097e539b821bba0a096384834e14188605d9c87e1c2a45e4bae8c1f07a7): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" podUID="084dcdaf-73d5-45a7-9d98-a1dbcd76db41" Dec 27 07:53:19 crc kubenswrapper[4934]: I1227 07:53:19.466953 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:19 crc kubenswrapper[4934]: I1227 07:53:19.470746 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:19 crc kubenswrapper[4934]: E1227 07:53:19.500350 4934 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-tx6z5_openshift-operators_f97e8d86-1e9e-4c5a-a164-2600193c576b_0(ebda49f5e04a06eef9816dbb3c5fc225793e2d2ef8ee0643322dad97f21218a8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 27 07:53:19 crc kubenswrapper[4934]: E1227 07:53:19.500595 4934 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-tx6z5_openshift-operators_f97e8d86-1e9e-4c5a-a164-2600193c576b_0(ebda49f5e04a06eef9816dbb3c5fc225793e2d2ef8ee0643322dad97f21218a8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:19 crc kubenswrapper[4934]: E1227 07:53:19.500619 4934 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-tx6z5_openshift-operators_f97e8d86-1e9e-4c5a-a164-2600193c576b_0(ebda49f5e04a06eef9816dbb3c5fc225793e2d2ef8ee0643322dad97f21218a8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:19 crc kubenswrapper[4934]: E1227 07:53:19.500659 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5bf474d74f-tx6z5_openshift-operators(f97e8d86-1e9e-4c5a-a164-2600193c576b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5bf474d74f-tx6z5_openshift-operators(f97e8d86-1e9e-4c5a-a164-2600193c576b)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-tx6z5_openshift-operators_f97e8d86-1e9e-4c5a-a164-2600193c576b_0(ebda49f5e04a06eef9816dbb3c5fc225793e2d2ef8ee0643322dad97f21218a8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" podUID="f97e8d86-1e9e-4c5a-a164-2600193c576b" Dec 27 07:53:20 crc kubenswrapper[4934]: I1227 07:53:20.467297 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" Dec 27 07:53:20 crc kubenswrapper[4934]: I1227 07:53:20.467304 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:20 crc kubenswrapper[4934]: I1227 07:53:20.467789 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:20 crc kubenswrapper[4934]: I1227 07:53:20.467826 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" Dec 27 07:53:20 crc kubenswrapper[4934]: E1227 07:53:20.524304 4934 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-lnknm_openshift-operators_4575d285-879f-4cc3-a9b9-406dc96e852d_0(46efcd697b6f426a334fb75e1edc721c8ea2d727a356bcfd5229a73511dd9665): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 27 07:53:20 crc kubenswrapper[4934]: E1227 07:53:20.524478 4934 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-lnknm_openshift-operators_4575d285-879f-4cc3-a9b9-406dc96e852d_0(46efcd697b6f426a334fb75e1edc721c8ea2d727a356bcfd5229a73511dd9665): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:20 crc kubenswrapper[4934]: E1227 07:53:20.524577 4934 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-lnknm_openshift-operators_4575d285-879f-4cc3-a9b9-406dc96e852d_0(46efcd697b6f426a334fb75e1edc721c8ea2d727a356bcfd5229a73511dd9665): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:20 crc kubenswrapper[4934]: E1227 07:53:20.524698 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-59bdc8b94-lnknm_openshift-operators(4575d285-879f-4cc3-a9b9-406dc96e852d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-59bdc8b94-lnknm_openshift-operators(4575d285-879f-4cc3-a9b9-406dc96e852d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-lnknm_openshift-operators_4575d285-879f-4cc3-a9b9-406dc96e852d_0(46efcd697b6f426a334fb75e1edc721c8ea2d727a356bcfd5229a73511dd9665): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" podUID="4575d285-879f-4cc3-a9b9-406dc96e852d" Dec 27 07:53:20 crc kubenswrapper[4934]: E1227 07:53:20.529218 4934 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-8llj6_openshift-operators_be52e823-2384-41da-b213-85e7fe61250e_0(916cbfb6aad9efb62432a05c7563533f8fcb35a1e586ab00e7d5306f8d8bb2d8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 27 07:53:20 crc kubenswrapper[4934]: E1227 07:53:20.529301 4934 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-8llj6_openshift-operators_be52e823-2384-41da-b213-85e7fe61250e_0(916cbfb6aad9efb62432a05c7563533f8fcb35a1e586ab00e7d5306f8d8bb2d8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" Dec 27 07:53:20 crc kubenswrapper[4934]: E1227 07:53:20.529363 4934 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-8llj6_openshift-operators_be52e823-2384-41da-b213-85e7fe61250e_0(916cbfb6aad9efb62432a05c7563533f8fcb35a1e586ab00e7d5306f8d8bb2d8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" Dec 27 07:53:20 crc kubenswrapper[4934]: E1227 07:53:20.529436 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-68bc856cb9-8llj6_openshift-operators(be52e823-2384-41da-b213-85e7fe61250e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-68bc856cb9-8llj6_openshift-operators(be52e823-2384-41da-b213-85e7fe61250e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-8llj6_openshift-operators_be52e823-2384-41da-b213-85e7fe61250e_0(916cbfb6aad9efb62432a05c7563533f8fcb35a1e586ab00e7d5306f8d8bb2d8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" podUID="be52e823-2384-41da-b213-85e7fe61250e" Dec 27 07:53:23 crc kubenswrapper[4934]: I1227 07:53:23.468242 4934 scope.go:117] "RemoveContainer" containerID="b3182ba625b9908dd1c4de2aa6395b330f542942567d562205e6ee11080b342a" Dec 27 07:53:24 crc kubenswrapper[4934]: I1227 07:53:24.819354 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-jzv2s_1bdd4b78-f615-4771-9448-8faa03dbc4f5/kube-multus/2.log" Dec 27 07:53:24 crc kubenswrapper[4934]: I1227 07:53:24.819673 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-jzv2s" event={"ID":"1bdd4b78-f615-4771-9448-8faa03dbc4f5","Type":"ContainerStarted","Data":"34972daa0b680888f0f86a5c1ee063c39d499c5490ff3e0e9fa0f3feaa96373e"} Dec 27 07:53:28 crc kubenswrapper[4934]: I1227 07:53:28.228273 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qbcrd" Dec 27 07:53:28 crc kubenswrapper[4934]: I1227 07:53:28.467371 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" Dec 27 07:53:28 crc kubenswrapper[4934]: I1227 07:53:28.468149 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" Dec 27 07:53:28 crc kubenswrapper[4934]: I1227 07:53:28.670400 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn"] Dec 27 07:53:28 crc kubenswrapper[4934]: I1227 07:53:28.843965 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" event={"ID":"b86f501c-2f1d-4131-820b-c9b71c4973dc","Type":"ContainerStarted","Data":"d4f933eb75867a03a5b3872b86b1601bb5a38aed572ea18280e21731ece19a49"} Dec 27 07:53:30 crc kubenswrapper[4934]: I1227 07:53:30.466916 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:30 crc kubenswrapper[4934]: I1227 07:53:30.467689 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:30 crc kubenswrapper[4934]: I1227 07:53:30.902359 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-tx6z5"] Dec 27 07:53:30 crc kubenswrapper[4934]: W1227 07:53:30.917903 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf97e8d86_1e9e_4c5a_a164_2600193c576b.slice/crio-0b650edf5dffe2a0affdd1f9f72137ab303a1ba3678d0666cc301504e0dbcb00 WatchSource:0}: Error finding container 0b650edf5dffe2a0affdd1f9f72137ab303a1ba3678d0666cc301504e0dbcb00: Status 404 returned error can't find the container with id 0b650edf5dffe2a0affdd1f9f72137ab303a1ba3678d0666cc301504e0dbcb00 Dec 27 07:53:31 crc kubenswrapper[4934]: I1227 07:53:31.862934 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" event={"ID":"f97e8d86-1e9e-4c5a-a164-2600193c576b","Type":"ContainerStarted","Data":"0b650edf5dffe2a0affdd1f9f72137ab303a1ba3678d0666cc301504e0dbcb00"} Dec 27 07:53:32 crc kubenswrapper[4934]: I1227 07:53:32.466619 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" Dec 27 07:53:32 crc kubenswrapper[4934]: I1227 07:53:32.467786 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" Dec 27 07:53:34 crc kubenswrapper[4934]: I1227 07:53:34.467326 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" Dec 27 07:53:34 crc kubenswrapper[4934]: I1227 07:53:34.468375 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" Dec 27 07:53:34 crc kubenswrapper[4934]: I1227 07:53:34.467358 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:34 crc kubenswrapper[4934]: I1227 07:53:34.469612 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:34 crc kubenswrapper[4934]: I1227 07:53:34.628732 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8"] Dec 27 07:53:34 crc kubenswrapper[4934]: W1227 07:53:34.646749 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod084dcdaf_73d5_45a7_9d98_a1dbcd76db41.slice/crio-67fdce65e212cca72899efc0f00c6db7472d6a9575e7cd6524bb5a12422eee99 WatchSource:0}: Error finding container 67fdce65e212cca72899efc0f00c6db7472d6a9575e7cd6524bb5a12422eee99: Status 404 returned error can't find the container with id 67fdce65e212cca72899efc0f00c6db7472d6a9575e7cd6524bb5a12422eee99 Dec 27 07:53:34 crc kubenswrapper[4934]: I1227 07:53:34.778361 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-lnknm"] Dec 27 07:53:34 crc kubenswrapper[4934]: W1227 07:53:34.778599 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4575d285_879f_4cc3_a9b9_406dc96e852d.slice/crio-08172b04f5481541f922733526f745271a9d41e03ba66356bc31221ef0b8300c WatchSource:0}: Error finding container 08172b04f5481541f922733526f745271a9d41e03ba66356bc31221ef0b8300c: Status 404 returned error can't find the container with id 08172b04f5481541f922733526f745271a9d41e03ba66356bc31221ef0b8300c Dec 27 07:53:34 crc kubenswrapper[4934]: I1227 07:53:34.881905 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" event={"ID":"4575d285-879f-4cc3-a9b9-406dc96e852d","Type":"ContainerStarted","Data":"08172b04f5481541f922733526f745271a9d41e03ba66356bc31221ef0b8300c"} Dec 27 07:53:34 crc kubenswrapper[4934]: I1227 07:53:34.883155 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" event={"ID":"b86f501c-2f1d-4131-820b-c9b71c4973dc","Type":"ContainerStarted","Data":"23f5a8bbcb8bb85c26421f6170d0ccb0b414b8a383327ad3d4e43aaa50e034a6"} Dec 27 07:53:34 crc kubenswrapper[4934]: I1227 07:53:34.884571 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" event={"ID":"084dcdaf-73d5-45a7-9d98-a1dbcd76db41","Type":"ContainerStarted","Data":"70d7f1a5e9bf99f4da2aeffd376ca9a4378b76aef6fa8fee97cc9d1913ceee30"} Dec 27 07:53:34 crc kubenswrapper[4934]: I1227 07:53:34.884590 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" event={"ID":"084dcdaf-73d5-45a7-9d98-a1dbcd76db41","Type":"ContainerStarted","Data":"67fdce65e212cca72899efc0f00c6db7472d6a9575e7cd6524bb5a12422eee99"} Dec 27 07:53:34 crc kubenswrapper[4934]: I1227 07:53:34.913849 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn" podStartSLOduration=25.177160107 podStartE2EDuration="30.913823481s" podCreationTimestamp="2025-12-27 07:53:04 +0000 UTC" firstStartedPulling="2025-12-27 07:53:28.683909935 +0000 UTC m=+669.504350529" lastFinishedPulling="2025-12-27 07:53:34.420573279 +0000 UTC m=+675.241013903" observedRunningTime="2025-12-27 07:53:34.901393905 +0000 UTC m=+675.721834499" watchObservedRunningTime="2025-12-27 07:53:34.913823481 +0000 UTC m=+675.734264075" Dec 27 07:53:34 crc kubenswrapper[4934]: I1227 07:53:34.988786 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8" podStartSLOduration=30.988761995 podStartE2EDuration="30.988761995s" podCreationTimestamp="2025-12-27 07:53:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:53:34.924228045 +0000 UTC m=+675.744668639" watchObservedRunningTime="2025-12-27 07:53:34.988761995 +0000 UTC m=+675.809202589" Dec 27 07:53:34 crc kubenswrapper[4934]: I1227 07:53:34.989794 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6"] Dec 27 07:53:35 crc kubenswrapper[4934]: I1227 07:53:35.894009 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" event={"ID":"be52e823-2384-41da-b213-85e7fe61250e","Type":"ContainerStarted","Data":"f5d59f7b533490f7e1ce069df36d27fc2a793d268b965f10175b6bd58062b979"} Dec 27 07:53:37 crc kubenswrapper[4934]: I1227 07:53:37.912628 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" event={"ID":"f97e8d86-1e9e-4c5a-a164-2600193c576b","Type":"ContainerStarted","Data":"fbd12ece2209c9dead9cee738d2f581fe32ccca160efd4cf58b4a611c9ce0535"} Dec 27 07:53:37 crc kubenswrapper[4934]: I1227 07:53:37.913203 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:37 crc kubenswrapper[4934]: I1227 07:53:37.933297 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" podStartSLOduration=27.796705313 podStartE2EDuration="33.933278578s" podCreationTimestamp="2025-12-27 07:53:04 +0000 UTC" firstStartedPulling="2025-12-27 07:53:30.921009643 +0000 UTC m=+671.741450237" lastFinishedPulling="2025-12-27 07:53:37.057582908 +0000 UTC m=+677.878023502" observedRunningTime="2025-12-27 07:53:37.930164419 +0000 UTC m=+678.750605023" watchObservedRunningTime="2025-12-27 07:53:37.933278578 +0000 UTC m=+678.753719172" Dec 27 07:53:40 crc kubenswrapper[4934]: I1227 07:53:40.340105 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" event={"ID":"be52e823-2384-41da-b213-85e7fe61250e","Type":"ContainerStarted","Data":"2723b6dc0a64683608e30dbbfd73b7a91602bafceb5d6d5bbc111ef4f1b7124c"} Dec 27 07:53:42 crc kubenswrapper[4934]: I1227 07:53:42.355045 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" event={"ID":"4575d285-879f-4cc3-a9b9-406dc96e852d","Type":"ContainerStarted","Data":"0e9c0d7603353e305c0f51f83e3c3c23c80d0364f79fc2f47a4a05801de2061b"} Dec 27 07:53:42 crc kubenswrapper[4934]: I1227 07:53:42.355419 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:42 crc kubenswrapper[4934]: I1227 07:53:42.358042 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 07:53:42 crc kubenswrapper[4934]: I1227 07:53:42.380366 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" podStartSLOduration=31.304397384 podStartE2EDuration="38.380342316s" podCreationTimestamp="2025-12-27 07:53:04 +0000 UTC" firstStartedPulling="2025-12-27 07:53:34.782935625 +0000 UTC m=+675.603376219" lastFinishedPulling="2025-12-27 07:53:41.858880537 +0000 UTC m=+682.679321151" observedRunningTime="2025-12-27 07:53:42.377536145 +0000 UTC m=+683.197976759" watchObservedRunningTime="2025-12-27 07:53:42.380342316 +0000 UTC m=+683.200782930" Dec 27 07:53:42 crc kubenswrapper[4934]: I1227 07:53:42.384369 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-8llj6" podStartSLOduration=35.859833789 podStartE2EDuration="39.384352018s" podCreationTimestamp="2025-12-27 07:53:03 +0000 UTC" firstStartedPulling="2025-12-27 07:53:35.001854258 +0000 UTC m=+675.822294852" lastFinishedPulling="2025-12-27 07:53:38.526372467 +0000 UTC m=+679.346813081" observedRunningTime="2025-12-27 07:53:40.358134226 +0000 UTC m=+681.178574860" watchObservedRunningTime="2025-12-27 07:53:42.384352018 +0000 UTC m=+683.204792632" Dec 27 07:53:44 crc kubenswrapper[4934]: I1227 07:53:44.837147 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 07:53:48 crc kubenswrapper[4934]: I1227 07:53:48.944448 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-hmzlk"] Dec 27 07:53:48 crc kubenswrapper[4934]: I1227 07:53:48.945776 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-hmzlk" Dec 27 07:53:48 crc kubenswrapper[4934]: I1227 07:53:48.948653 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-26fts"] Dec 27 07:53:48 crc kubenswrapper[4934]: I1227 07:53:48.949321 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-26fts" Dec 27 07:53:48 crc kubenswrapper[4934]: I1227 07:53:48.950816 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 27 07:53:48 crc kubenswrapper[4934]: I1227 07:53:48.951449 4934 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-cw5b2" Dec 27 07:53:48 crc kubenswrapper[4934]: I1227 07:53:48.951500 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 27 07:53:48 crc kubenswrapper[4934]: I1227 07:53:48.951324 4934 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-n6p98" Dec 27 07:53:48 crc kubenswrapper[4934]: I1227 07:53:48.958496 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-26fts"] Dec 27 07:53:48 crc kubenswrapper[4934]: I1227 07:53:48.962838 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-hmzlk"] Dec 27 07:53:48 crc kubenswrapper[4934]: I1227 07:53:48.976597 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-5kwg4"] Dec 27 07:53:48 crc kubenswrapper[4934]: I1227 07:53:48.977601 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-5kwg4" Dec 27 07:53:48 crc kubenswrapper[4934]: I1227 07:53:48.979368 4934 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-m4bvq" Dec 27 07:53:48 crc kubenswrapper[4934]: I1227 07:53:48.992075 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5kmh\" (UniqueName: \"kubernetes.io/projected/dab3e53b-cbb2-4423-a8e3-432972ecedc3-kube-api-access-k5kmh\") pod \"cert-manager-webhook-687f57d79b-5kwg4\" (UID: \"dab3e53b-cbb2-4423-a8e3-432972ecedc3\") " pod="cert-manager/cert-manager-webhook-687f57d79b-5kwg4" Dec 27 07:53:48 crc kubenswrapper[4934]: I1227 07:53:48.992272 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srnkb\" (UniqueName: \"kubernetes.io/projected/1c38c7d2-7994-4d2b-83a2-d98066e4ba5b-kube-api-access-srnkb\") pod \"cert-manager-858654f9db-26fts\" (UID: \"1c38c7d2-7994-4d2b-83a2-d98066e4ba5b\") " pod="cert-manager/cert-manager-858654f9db-26fts" Dec 27 07:53:48 crc kubenswrapper[4934]: I1227 07:53:48.992501 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwhc2\" (UniqueName: \"kubernetes.io/projected/e14d5680-64cf-4640-a505-62020b9c31f4-kube-api-access-lwhc2\") pod \"cert-manager-cainjector-cf98fcc89-hmzlk\" (UID: \"e14d5680-64cf-4640-a505-62020b9c31f4\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-hmzlk" Dec 27 07:53:48 crc kubenswrapper[4934]: I1227 07:53:48.996154 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-5kwg4"] Dec 27 07:53:49 crc kubenswrapper[4934]: I1227 07:53:49.093336 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srnkb\" (UniqueName: \"kubernetes.io/projected/1c38c7d2-7994-4d2b-83a2-d98066e4ba5b-kube-api-access-srnkb\") pod \"cert-manager-858654f9db-26fts\" (UID: \"1c38c7d2-7994-4d2b-83a2-d98066e4ba5b\") " pod="cert-manager/cert-manager-858654f9db-26fts" Dec 27 07:53:49 crc kubenswrapper[4934]: I1227 07:53:49.093405 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwhc2\" (UniqueName: \"kubernetes.io/projected/e14d5680-64cf-4640-a505-62020b9c31f4-kube-api-access-lwhc2\") pod \"cert-manager-cainjector-cf98fcc89-hmzlk\" (UID: \"e14d5680-64cf-4640-a505-62020b9c31f4\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-hmzlk" Dec 27 07:53:49 crc kubenswrapper[4934]: I1227 07:53:49.093470 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5kmh\" (UniqueName: \"kubernetes.io/projected/dab3e53b-cbb2-4423-a8e3-432972ecedc3-kube-api-access-k5kmh\") pod \"cert-manager-webhook-687f57d79b-5kwg4\" (UID: \"dab3e53b-cbb2-4423-a8e3-432972ecedc3\") " pod="cert-manager/cert-manager-webhook-687f57d79b-5kwg4" Dec 27 07:53:49 crc kubenswrapper[4934]: I1227 07:53:49.112346 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5kmh\" (UniqueName: \"kubernetes.io/projected/dab3e53b-cbb2-4423-a8e3-432972ecedc3-kube-api-access-k5kmh\") pod \"cert-manager-webhook-687f57d79b-5kwg4\" (UID: \"dab3e53b-cbb2-4423-a8e3-432972ecedc3\") " pod="cert-manager/cert-manager-webhook-687f57d79b-5kwg4" Dec 27 07:53:49 crc kubenswrapper[4934]: I1227 07:53:49.114152 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwhc2\" (UniqueName: \"kubernetes.io/projected/e14d5680-64cf-4640-a505-62020b9c31f4-kube-api-access-lwhc2\") pod \"cert-manager-cainjector-cf98fcc89-hmzlk\" (UID: \"e14d5680-64cf-4640-a505-62020b9c31f4\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-hmzlk" Dec 27 07:53:49 crc kubenswrapper[4934]: I1227 07:53:49.114236 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srnkb\" (UniqueName: \"kubernetes.io/projected/1c38c7d2-7994-4d2b-83a2-d98066e4ba5b-kube-api-access-srnkb\") pod \"cert-manager-858654f9db-26fts\" (UID: \"1c38c7d2-7994-4d2b-83a2-d98066e4ba5b\") " pod="cert-manager/cert-manager-858654f9db-26fts" Dec 27 07:53:49 crc kubenswrapper[4934]: I1227 07:53:49.262247 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-hmzlk" Dec 27 07:53:49 crc kubenswrapper[4934]: I1227 07:53:49.277424 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-26fts" Dec 27 07:53:49 crc kubenswrapper[4934]: I1227 07:53:49.289423 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-5kwg4" Dec 27 07:53:49 crc kubenswrapper[4934]: I1227 07:53:49.630847 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-hmzlk"] Dec 27 07:53:49 crc kubenswrapper[4934]: I1227 07:53:49.646361 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-26fts"] Dec 27 07:53:49 crc kubenswrapper[4934]: I1227 07:53:49.687811 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-5kwg4"] Dec 27 07:53:50 crc kubenswrapper[4934]: I1227 07:53:50.430464 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-hmzlk" event={"ID":"e14d5680-64cf-4640-a505-62020b9c31f4","Type":"ContainerStarted","Data":"816f87528f0e7159332b7848f3955699bf9ea4bfe3c6b48ea8a5d4c09c03763d"} Dec 27 07:53:50 crc kubenswrapper[4934]: I1227 07:53:50.432205 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-5kwg4" event={"ID":"dab3e53b-cbb2-4423-a8e3-432972ecedc3","Type":"ContainerStarted","Data":"5bd85d7cbe5596312d066452d22de7a43f152beba0a21370bca02168744e700f"} Dec 27 07:53:50 crc kubenswrapper[4934]: I1227 07:53:50.433816 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-26fts" event={"ID":"1c38c7d2-7994-4d2b-83a2-d98066e4ba5b","Type":"ContainerStarted","Data":"265d85371357c03cf8e160bd038968115bdf6c9c071e70aae65f014387403573"} Dec 27 07:53:54 crc kubenswrapper[4934]: I1227 07:53:54.478701 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-hmzlk" event={"ID":"e14d5680-64cf-4640-a505-62020b9c31f4","Type":"ContainerStarted","Data":"54e1cb84f808ecabd8075d1d442bdf7b4b198ea3cecca510d9f9bbdc06d01195"} Dec 27 07:53:54 crc kubenswrapper[4934]: I1227 07:53:54.482633 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-5kwg4" event={"ID":"dab3e53b-cbb2-4423-a8e3-432972ecedc3","Type":"ContainerStarted","Data":"c2917505796f55bde5592ccf2bb65452f4640493c5846e3d72b01a05e2fb8bce"} Dec 27 07:53:54 crc kubenswrapper[4934]: I1227 07:53:54.482745 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-5kwg4" Dec 27 07:53:54 crc kubenswrapper[4934]: I1227 07:53:54.485714 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-26fts" event={"ID":"1c38c7d2-7994-4d2b-83a2-d98066e4ba5b","Type":"ContainerStarted","Data":"7677be7d7103497ab6bb79b2e35756169f587a7b7fa6f3e9a5f23dbea3f8ba68"} Dec 27 07:53:54 crc kubenswrapper[4934]: I1227 07:53:54.506860 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-hmzlk" podStartSLOduration=2.379023325 podStartE2EDuration="6.506838232s" podCreationTimestamp="2025-12-27 07:53:48 +0000 UTC" firstStartedPulling="2025-12-27 07:53:49.647620051 +0000 UTC m=+690.468060645" lastFinishedPulling="2025-12-27 07:53:53.775434948 +0000 UTC m=+694.595875552" observedRunningTime="2025-12-27 07:53:54.498866559 +0000 UTC m=+695.319307213" watchObservedRunningTime="2025-12-27 07:53:54.506838232 +0000 UTC m=+695.327278856" Dec 27 07:53:54 crc kubenswrapper[4934]: I1227 07:53:54.536230 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-5kwg4" podStartSLOduration=2.388977447 podStartE2EDuration="6.536195158s" podCreationTimestamp="2025-12-27 07:53:48 +0000 UTC" firstStartedPulling="2025-12-27 07:53:49.706476276 +0000 UTC m=+690.526916870" lastFinishedPulling="2025-12-27 07:53:53.853693987 +0000 UTC m=+694.674134581" observedRunningTime="2025-12-27 07:53:54.534562536 +0000 UTC m=+695.355003150" watchObservedRunningTime="2025-12-27 07:53:54.536195158 +0000 UTC m=+695.356635792" Dec 27 07:53:54 crc kubenswrapper[4934]: I1227 07:53:54.555563 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-26fts" podStartSLOduration=2.45402943 podStartE2EDuration="6.555492238s" podCreationTimestamp="2025-12-27 07:53:48 +0000 UTC" firstStartedPulling="2025-12-27 07:53:49.675369186 +0000 UTC m=+690.495809780" lastFinishedPulling="2025-12-27 07:53:53.776831984 +0000 UTC m=+694.597272588" observedRunningTime="2025-12-27 07:53:54.554641156 +0000 UTC m=+695.375081790" watchObservedRunningTime="2025-12-27 07:53:54.555492238 +0000 UTC m=+695.375932842" Dec 27 07:53:59 crc kubenswrapper[4934]: I1227 07:53:59.292024 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-5kwg4" Dec 27 07:54:15 crc kubenswrapper[4934]: I1227 07:54:15.330485 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 07:54:15 crc kubenswrapper[4934]: I1227 07:54:15.331351 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.011861 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8"] Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.014025 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.016497 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.024208 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8"] Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.038176 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/74e9f2f0-b2fa-4574-9378-263ef6468eda-util\") pod \"40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8\" (UID: \"74e9f2f0-b2fa-4574-9378-263ef6468eda\") " pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.038472 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f662m\" (UniqueName: \"kubernetes.io/projected/74e9f2f0-b2fa-4574-9378-263ef6468eda-kube-api-access-f662m\") pod \"40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8\" (UID: \"74e9f2f0-b2fa-4574-9378-263ef6468eda\") " pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.038543 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/74e9f2f0-b2fa-4574-9378-263ef6468eda-bundle\") pod \"40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8\" (UID: \"74e9f2f0-b2fa-4574-9378-263ef6468eda\") " pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.140255 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f662m\" (UniqueName: \"kubernetes.io/projected/74e9f2f0-b2fa-4574-9378-263ef6468eda-kube-api-access-f662m\") pod \"40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8\" (UID: \"74e9f2f0-b2fa-4574-9378-263ef6468eda\") " pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.140321 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/74e9f2f0-b2fa-4574-9378-263ef6468eda-bundle\") pod \"40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8\" (UID: \"74e9f2f0-b2fa-4574-9378-263ef6468eda\") " pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.140375 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/74e9f2f0-b2fa-4574-9378-263ef6468eda-util\") pod \"40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8\" (UID: \"74e9f2f0-b2fa-4574-9378-263ef6468eda\") " pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.140982 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/74e9f2f0-b2fa-4574-9378-263ef6468eda-util\") pod \"40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8\" (UID: \"74e9f2f0-b2fa-4574-9378-263ef6468eda\") " pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.141260 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/74e9f2f0-b2fa-4574-9378-263ef6468eda-bundle\") pod \"40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8\" (UID: \"74e9f2f0-b2fa-4574-9378-263ef6468eda\") " pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.172464 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f662m\" (UniqueName: \"kubernetes.io/projected/74e9f2f0-b2fa-4574-9378-263ef6468eda-kube-api-access-f662m\") pod \"40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8\" (UID: \"74e9f2f0-b2fa-4574-9378-263ef6468eda\") " pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.203177 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt"] Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.204413 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.217046 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt"] Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.242178 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2befd093-f28c-4779-b290-eb99f3ef5cc2-bundle\") pod \"19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt\" (UID: \"2befd093-f28c-4779-b290-eb99f3ef5cc2\") " pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.242470 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2befd093-f28c-4779-b290-eb99f3ef5cc2-util\") pod \"19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt\" (UID: \"2befd093-f28c-4779-b290-eb99f3ef5cc2\") " pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.242738 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcntg\" (UniqueName: \"kubernetes.io/projected/2befd093-f28c-4779-b290-eb99f3ef5cc2-kube-api-access-jcntg\") pod \"19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt\" (UID: \"2befd093-f28c-4779-b290-eb99f3ef5cc2\") " pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.338854 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.344067 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2befd093-f28c-4779-b290-eb99f3ef5cc2-bundle\") pod \"19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt\" (UID: \"2befd093-f28c-4779-b290-eb99f3ef5cc2\") " pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.344127 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2befd093-f28c-4779-b290-eb99f3ef5cc2-util\") pod \"19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt\" (UID: \"2befd093-f28c-4779-b290-eb99f3ef5cc2\") " pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.344171 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcntg\" (UniqueName: \"kubernetes.io/projected/2befd093-f28c-4779-b290-eb99f3ef5cc2-kube-api-access-jcntg\") pod \"19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt\" (UID: \"2befd093-f28c-4779-b290-eb99f3ef5cc2\") " pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.344716 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2befd093-f28c-4779-b290-eb99f3ef5cc2-bundle\") pod \"19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt\" (UID: \"2befd093-f28c-4779-b290-eb99f3ef5cc2\") " pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.344787 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2befd093-f28c-4779-b290-eb99f3ef5cc2-util\") pod \"19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt\" (UID: \"2befd093-f28c-4779-b290-eb99f3ef5cc2\") " pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.365527 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcntg\" (UniqueName: \"kubernetes.io/projected/2befd093-f28c-4779-b290-eb99f3ef5cc2-kube-api-access-jcntg\") pod \"19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt\" (UID: \"2befd093-f28c-4779-b290-eb99f3ef5cc2\") " pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.523553 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt" Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.580667 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8"] Dec 27 07:54:22 crc kubenswrapper[4934]: W1227 07:54:22.589212 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74e9f2f0_b2fa_4574_9378_263ef6468eda.slice/crio-a6161cfab4f9076444d4eee40a9dcd48c7f10742253fd941ffa41bd62819d400 WatchSource:0}: Error finding container a6161cfab4f9076444d4eee40a9dcd48c7f10742253fd941ffa41bd62819d400: Status 404 returned error can't find the container with id a6161cfab4f9076444d4eee40a9dcd48c7f10742253fd941ffa41bd62819d400 Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.704202 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8" event={"ID":"74e9f2f0-b2fa-4574-9378-263ef6468eda","Type":"ContainerStarted","Data":"a6161cfab4f9076444d4eee40a9dcd48c7f10742253fd941ffa41bd62819d400"} Dec 27 07:54:22 crc kubenswrapper[4934]: I1227 07:54:22.948824 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt"] Dec 27 07:54:22 crc kubenswrapper[4934]: W1227 07:54:22.956285 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2befd093_f28c_4779_b290_eb99f3ef5cc2.slice/crio-7da6d949a4f173daf9949c86728aa288811bdd0f97d8b7d0be86bcc865d06ec8 WatchSource:0}: Error finding container 7da6d949a4f173daf9949c86728aa288811bdd0f97d8b7d0be86bcc865d06ec8: Status 404 returned error can't find the container with id 7da6d949a4f173daf9949c86728aa288811bdd0f97d8b7d0be86bcc865d06ec8 Dec 27 07:54:23 crc kubenswrapper[4934]: I1227 07:54:23.714747 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt" event={"ID":"2befd093-f28c-4779-b290-eb99f3ef5cc2","Type":"ContainerStarted","Data":"7da6d949a4f173daf9949c86728aa288811bdd0f97d8b7d0be86bcc865d06ec8"} Dec 27 07:54:26 crc kubenswrapper[4934]: I1227 07:54:26.737038 4934 generic.go:334] "Generic (PLEG): container finished" podID="74e9f2f0-b2fa-4574-9378-263ef6468eda" containerID="e2c16290370f5eba878e1bf2e6857093e28733ee90b55aef6d8d7f436a687c83" exitCode=0 Dec 27 07:54:26 crc kubenswrapper[4934]: I1227 07:54:26.737136 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8" event={"ID":"74e9f2f0-b2fa-4574-9378-263ef6468eda","Type":"ContainerDied","Data":"e2c16290370f5eba878e1bf2e6857093e28733ee90b55aef6d8d7f436a687c83"} Dec 27 07:54:26 crc kubenswrapper[4934]: I1227 07:54:26.745214 4934 generic.go:334] "Generic (PLEG): container finished" podID="2befd093-f28c-4779-b290-eb99f3ef5cc2" containerID="80e64b6fd85fbb354a8f1337b11f2e35f242e9bd684055921b74b05b694821c2" exitCode=0 Dec 27 07:54:26 crc kubenswrapper[4934]: I1227 07:54:26.745256 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt" event={"ID":"2befd093-f28c-4779-b290-eb99f3ef5cc2","Type":"ContainerDied","Data":"80e64b6fd85fbb354a8f1337b11f2e35f242e9bd684055921b74b05b694821c2"} Dec 27 07:54:28 crc kubenswrapper[4934]: I1227 07:54:28.763984 4934 generic.go:334] "Generic (PLEG): container finished" podID="2befd093-f28c-4779-b290-eb99f3ef5cc2" containerID="a6d04fa71af1d77b26a243e1401dd90207077b8ab482ba1acd5d5b1dbaf9e022" exitCode=0 Dec 27 07:54:28 crc kubenswrapper[4934]: I1227 07:54:28.764367 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt" event={"ID":"2befd093-f28c-4779-b290-eb99f3ef5cc2","Type":"ContainerDied","Data":"a6d04fa71af1d77b26a243e1401dd90207077b8ab482ba1acd5d5b1dbaf9e022"} Dec 27 07:54:28 crc kubenswrapper[4934]: I1227 07:54:28.768793 4934 generic.go:334] "Generic (PLEG): container finished" podID="74e9f2f0-b2fa-4574-9378-263ef6468eda" containerID="a2303e9ce226ec8508026ee48d7d8fa941042717b066b3b759818cb49325d3d4" exitCode=0 Dec 27 07:54:28 crc kubenswrapper[4934]: I1227 07:54:28.768841 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8" event={"ID":"74e9f2f0-b2fa-4574-9378-263ef6468eda","Type":"ContainerDied","Data":"a2303e9ce226ec8508026ee48d7d8fa941042717b066b3b759818cb49325d3d4"} Dec 27 07:54:29 crc kubenswrapper[4934]: I1227 07:54:29.781395 4934 generic.go:334] "Generic (PLEG): container finished" podID="74e9f2f0-b2fa-4574-9378-263ef6468eda" containerID="fe41cf42bb33a39a0b38bc9782c14bb69fe36977a904eaac8e8b4c2bcdfd530c" exitCode=0 Dec 27 07:54:29 crc kubenswrapper[4934]: I1227 07:54:29.781512 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8" event={"ID":"74e9f2f0-b2fa-4574-9378-263ef6468eda","Type":"ContainerDied","Data":"fe41cf42bb33a39a0b38bc9782c14bb69fe36977a904eaac8e8b4c2bcdfd530c"} Dec 27 07:54:29 crc kubenswrapper[4934]: I1227 07:54:29.785204 4934 generic.go:334] "Generic (PLEG): container finished" podID="2befd093-f28c-4779-b290-eb99f3ef5cc2" containerID="482c4e8456b4cc26757eb97324e48b85ef4c3b479e3724d99239762d427463dd" exitCode=0 Dec 27 07:54:29 crc kubenswrapper[4934]: I1227 07:54:29.785267 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt" event={"ID":"2befd093-f28c-4779-b290-eb99f3ef5cc2","Type":"ContainerDied","Data":"482c4e8456b4cc26757eb97324e48b85ef4c3b479e3724d99239762d427463dd"} Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.167644 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8" Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.174161 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt" Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.202127 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/74e9f2f0-b2fa-4574-9378-263ef6468eda-util\") pod \"74e9f2f0-b2fa-4574-9378-263ef6468eda\" (UID: \"74e9f2f0-b2fa-4574-9378-263ef6468eda\") " Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.202225 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jcntg\" (UniqueName: \"kubernetes.io/projected/2befd093-f28c-4779-b290-eb99f3ef5cc2-kube-api-access-jcntg\") pod \"2befd093-f28c-4779-b290-eb99f3ef5cc2\" (UID: \"2befd093-f28c-4779-b290-eb99f3ef5cc2\") " Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.202245 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2befd093-f28c-4779-b290-eb99f3ef5cc2-bundle\") pod \"2befd093-f28c-4779-b290-eb99f3ef5cc2\" (UID: \"2befd093-f28c-4779-b290-eb99f3ef5cc2\") " Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.202323 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2befd093-f28c-4779-b290-eb99f3ef5cc2-util\") pod \"2befd093-f28c-4779-b290-eb99f3ef5cc2\" (UID: \"2befd093-f28c-4779-b290-eb99f3ef5cc2\") " Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.202354 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/74e9f2f0-b2fa-4574-9378-263ef6468eda-bundle\") pod \"74e9f2f0-b2fa-4574-9378-263ef6468eda\" (UID: \"74e9f2f0-b2fa-4574-9378-263ef6468eda\") " Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.202392 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f662m\" (UniqueName: \"kubernetes.io/projected/74e9f2f0-b2fa-4574-9378-263ef6468eda-kube-api-access-f662m\") pod \"74e9f2f0-b2fa-4574-9378-263ef6468eda\" (UID: \"74e9f2f0-b2fa-4574-9378-263ef6468eda\") " Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.206337 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2befd093-f28c-4779-b290-eb99f3ef5cc2-bundle" (OuterVolumeSpecName: "bundle") pod "2befd093-f28c-4779-b290-eb99f3ef5cc2" (UID: "2befd093-f28c-4779-b290-eb99f3ef5cc2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.206991 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74e9f2f0-b2fa-4574-9378-263ef6468eda-bundle" (OuterVolumeSpecName: "bundle") pod "74e9f2f0-b2fa-4574-9378-263ef6468eda" (UID: "74e9f2f0-b2fa-4574-9378-263ef6468eda"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.209388 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74e9f2f0-b2fa-4574-9378-263ef6468eda-kube-api-access-f662m" (OuterVolumeSpecName: "kube-api-access-f662m") pod "74e9f2f0-b2fa-4574-9378-263ef6468eda" (UID: "74e9f2f0-b2fa-4574-9378-263ef6468eda"). InnerVolumeSpecName "kube-api-access-f662m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.218716 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74e9f2f0-b2fa-4574-9378-263ef6468eda-util" (OuterVolumeSpecName: "util") pod "74e9f2f0-b2fa-4574-9378-263ef6468eda" (UID: "74e9f2f0-b2fa-4574-9378-263ef6468eda"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.220145 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2befd093-f28c-4779-b290-eb99f3ef5cc2-util" (OuterVolumeSpecName: "util") pod "2befd093-f28c-4779-b290-eb99f3ef5cc2" (UID: "2befd093-f28c-4779-b290-eb99f3ef5cc2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.221045 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2befd093-f28c-4779-b290-eb99f3ef5cc2-kube-api-access-jcntg" (OuterVolumeSpecName: "kube-api-access-jcntg") pod "2befd093-f28c-4779-b290-eb99f3ef5cc2" (UID: "2befd093-f28c-4779-b290-eb99f3ef5cc2"). InnerVolumeSpecName "kube-api-access-jcntg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.304377 4934 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/74e9f2f0-b2fa-4574-9378-263ef6468eda-util\") on node \"crc\" DevicePath \"\"" Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.304425 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jcntg\" (UniqueName: \"kubernetes.io/projected/2befd093-f28c-4779-b290-eb99f3ef5cc2-kube-api-access-jcntg\") on node \"crc\" DevicePath \"\"" Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.304447 4934 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2befd093-f28c-4779-b290-eb99f3ef5cc2-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.304464 4934 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2befd093-f28c-4779-b290-eb99f3ef5cc2-util\") on node \"crc\" DevicePath \"\"" Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.304480 4934 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/74e9f2f0-b2fa-4574-9378-263ef6468eda-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.304537 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f662m\" (UniqueName: \"kubernetes.io/projected/74e9f2f0-b2fa-4574-9378-263ef6468eda-kube-api-access-f662m\") on node \"crc\" DevicePath \"\"" Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.814477 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8" event={"ID":"74e9f2f0-b2fa-4574-9378-263ef6468eda","Type":"ContainerDied","Data":"a6161cfab4f9076444d4eee40a9dcd48c7f10742253fd941ffa41bd62819d400"} Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.814532 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6161cfab4f9076444d4eee40a9dcd48c7f10742253fd941ffa41bd62819d400" Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.814636 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8" Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.823183 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt" event={"ID":"2befd093-f28c-4779-b290-eb99f3ef5cc2","Type":"ContainerDied","Data":"7da6d949a4f173daf9949c86728aa288811bdd0f97d8b7d0be86bcc865d06ec8"} Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.823282 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7da6d949a4f173daf9949c86728aa288811bdd0f97d8b7d0be86bcc865d06ec8" Dec 27 07:54:31 crc kubenswrapper[4934]: I1227 07:54:31.823303 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.083046 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9"] Dec 27 07:54:40 crc kubenswrapper[4934]: E1227 07:54:40.083714 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2befd093-f28c-4779-b290-eb99f3ef5cc2" containerName="pull" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.083726 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="2befd093-f28c-4779-b290-eb99f3ef5cc2" containerName="pull" Dec 27 07:54:40 crc kubenswrapper[4934]: E1227 07:54:40.083742 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2befd093-f28c-4779-b290-eb99f3ef5cc2" containerName="extract" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.083747 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="2befd093-f28c-4779-b290-eb99f3ef5cc2" containerName="extract" Dec 27 07:54:40 crc kubenswrapper[4934]: E1227 07:54:40.083755 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74e9f2f0-b2fa-4574-9378-263ef6468eda" containerName="extract" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.083762 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="74e9f2f0-b2fa-4574-9378-263ef6468eda" containerName="extract" Dec 27 07:54:40 crc kubenswrapper[4934]: E1227 07:54:40.083774 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74e9f2f0-b2fa-4574-9378-263ef6468eda" containerName="pull" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.083782 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="74e9f2f0-b2fa-4574-9378-263ef6468eda" containerName="pull" Dec 27 07:54:40 crc kubenswrapper[4934]: E1227 07:54:40.083794 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74e9f2f0-b2fa-4574-9378-263ef6468eda" containerName="util" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.083799 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="74e9f2f0-b2fa-4574-9378-263ef6468eda" containerName="util" Dec 27 07:54:40 crc kubenswrapper[4934]: E1227 07:54:40.083811 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2befd093-f28c-4779-b290-eb99f3ef5cc2" containerName="util" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.083816 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="2befd093-f28c-4779-b290-eb99f3ef5cc2" containerName="util" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.083911 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="2befd093-f28c-4779-b290-eb99f3ef5cc2" containerName="extract" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.083922 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="74e9f2f0-b2fa-4574-9378-263ef6468eda" containerName="extract" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.084524 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.087351 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-lzbrg" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.087479 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.087662 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.088683 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.088754 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.090030 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.104461 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9"] Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.134030 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plrsp\" (UniqueName: \"kubernetes.io/projected/00799e47-9b94-49bf-8bd1-d6bb5036285a-kube-api-access-plrsp\") pod \"loki-operator-controller-manager-77c7786d5-gz9d9\" (UID: \"00799e47-9b94-49bf-8bd1-d6bb5036285a\") " pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.134100 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/00799e47-9b94-49bf-8bd1-d6bb5036285a-manager-config\") pod \"loki-operator-controller-manager-77c7786d5-gz9d9\" (UID: \"00799e47-9b94-49bf-8bd1-d6bb5036285a\") " pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.134274 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/00799e47-9b94-49bf-8bd1-d6bb5036285a-webhook-cert\") pod \"loki-operator-controller-manager-77c7786d5-gz9d9\" (UID: \"00799e47-9b94-49bf-8bd1-d6bb5036285a\") " pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.134366 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/00799e47-9b94-49bf-8bd1-d6bb5036285a-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-77c7786d5-gz9d9\" (UID: \"00799e47-9b94-49bf-8bd1-d6bb5036285a\") " pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.134431 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/00799e47-9b94-49bf-8bd1-d6bb5036285a-apiservice-cert\") pod \"loki-operator-controller-manager-77c7786d5-gz9d9\" (UID: \"00799e47-9b94-49bf-8bd1-d6bb5036285a\") " pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.235573 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/00799e47-9b94-49bf-8bd1-d6bb5036285a-webhook-cert\") pod \"loki-operator-controller-manager-77c7786d5-gz9d9\" (UID: \"00799e47-9b94-49bf-8bd1-d6bb5036285a\") " pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.235629 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/00799e47-9b94-49bf-8bd1-d6bb5036285a-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-77c7786d5-gz9d9\" (UID: \"00799e47-9b94-49bf-8bd1-d6bb5036285a\") " pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.235655 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/00799e47-9b94-49bf-8bd1-d6bb5036285a-apiservice-cert\") pod \"loki-operator-controller-manager-77c7786d5-gz9d9\" (UID: \"00799e47-9b94-49bf-8bd1-d6bb5036285a\") " pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.235696 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plrsp\" (UniqueName: \"kubernetes.io/projected/00799e47-9b94-49bf-8bd1-d6bb5036285a-kube-api-access-plrsp\") pod \"loki-operator-controller-manager-77c7786d5-gz9d9\" (UID: \"00799e47-9b94-49bf-8bd1-d6bb5036285a\") " pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.235728 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/00799e47-9b94-49bf-8bd1-d6bb5036285a-manager-config\") pod \"loki-operator-controller-manager-77c7786d5-gz9d9\" (UID: \"00799e47-9b94-49bf-8bd1-d6bb5036285a\") " pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.236981 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/00799e47-9b94-49bf-8bd1-d6bb5036285a-manager-config\") pod \"loki-operator-controller-manager-77c7786d5-gz9d9\" (UID: \"00799e47-9b94-49bf-8bd1-d6bb5036285a\") " pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.240394 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/00799e47-9b94-49bf-8bd1-d6bb5036285a-webhook-cert\") pod \"loki-operator-controller-manager-77c7786d5-gz9d9\" (UID: \"00799e47-9b94-49bf-8bd1-d6bb5036285a\") " pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.240394 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/00799e47-9b94-49bf-8bd1-d6bb5036285a-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-77c7786d5-gz9d9\" (UID: \"00799e47-9b94-49bf-8bd1-d6bb5036285a\") " pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.247863 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/00799e47-9b94-49bf-8bd1-d6bb5036285a-apiservice-cert\") pod \"loki-operator-controller-manager-77c7786d5-gz9d9\" (UID: \"00799e47-9b94-49bf-8bd1-d6bb5036285a\") " pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.256873 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plrsp\" (UniqueName: \"kubernetes.io/projected/00799e47-9b94-49bf-8bd1-d6bb5036285a-kube-api-access-plrsp\") pod \"loki-operator-controller-manager-77c7786d5-gz9d9\" (UID: \"00799e47-9b94-49bf-8bd1-d6bb5036285a\") " pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.399908 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.594984 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9"] Dec 27 07:54:40 crc kubenswrapper[4934]: W1227 07:54:40.602037 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00799e47_9b94_49bf_8bd1_d6bb5036285a.slice/crio-63346f1e1426a768666e4bec086ed3de16e9d16c630e94366895ed026be358da WatchSource:0}: Error finding container 63346f1e1426a768666e4bec086ed3de16e9d16c630e94366895ed026be358da: Status 404 returned error can't find the container with id 63346f1e1426a768666e4bec086ed3de16e9d16c630e94366895ed026be358da Dec 27 07:54:40 crc kubenswrapper[4934]: I1227 07:54:40.882645 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" event={"ID":"00799e47-9b94-49bf-8bd1-d6bb5036285a","Type":"ContainerStarted","Data":"63346f1e1426a768666e4bec086ed3de16e9d16c630e94366895ed026be358da"} Dec 27 07:54:41 crc kubenswrapper[4934]: I1227 07:54:41.156582 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/cluster-logging-operator-79cf69ddc8-h47cv"] Dec 27 07:54:41 crc kubenswrapper[4934]: I1227 07:54:41.158353 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-79cf69ddc8-h47cv" Dec 27 07:54:41 crc kubenswrapper[4934]: I1227 07:54:41.160282 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"cluster-logging-operator-dockercfg-5ckrc" Dec 27 07:54:41 crc kubenswrapper[4934]: I1227 07:54:41.160838 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"openshift-service-ca.crt" Dec 27 07:54:41 crc kubenswrapper[4934]: I1227 07:54:41.161803 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"kube-root-ca.crt" Dec 27 07:54:41 crc kubenswrapper[4934]: I1227 07:54:41.172327 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-79cf69ddc8-h47cv"] Dec 27 07:54:41 crc kubenswrapper[4934]: I1227 07:54:41.249993 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrrsf\" (UniqueName: \"kubernetes.io/projected/8b6af0c6-05e2-4adc-8e98-7d2e4f6254ea-kube-api-access-lrrsf\") pod \"cluster-logging-operator-79cf69ddc8-h47cv\" (UID: \"8b6af0c6-05e2-4adc-8e98-7d2e4f6254ea\") " pod="openshift-logging/cluster-logging-operator-79cf69ddc8-h47cv" Dec 27 07:54:41 crc kubenswrapper[4934]: I1227 07:54:41.351797 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrrsf\" (UniqueName: \"kubernetes.io/projected/8b6af0c6-05e2-4adc-8e98-7d2e4f6254ea-kube-api-access-lrrsf\") pod \"cluster-logging-operator-79cf69ddc8-h47cv\" (UID: \"8b6af0c6-05e2-4adc-8e98-7d2e4f6254ea\") " pod="openshift-logging/cluster-logging-operator-79cf69ddc8-h47cv" Dec 27 07:54:41 crc kubenswrapper[4934]: I1227 07:54:41.393142 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrrsf\" (UniqueName: \"kubernetes.io/projected/8b6af0c6-05e2-4adc-8e98-7d2e4f6254ea-kube-api-access-lrrsf\") pod \"cluster-logging-operator-79cf69ddc8-h47cv\" (UID: \"8b6af0c6-05e2-4adc-8e98-7d2e4f6254ea\") " pod="openshift-logging/cluster-logging-operator-79cf69ddc8-h47cv" Dec 27 07:54:41 crc kubenswrapper[4934]: I1227 07:54:41.490194 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-79cf69ddc8-h47cv" Dec 27 07:54:41 crc kubenswrapper[4934]: I1227 07:54:41.818641 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-79cf69ddc8-h47cv"] Dec 27 07:54:41 crc kubenswrapper[4934]: W1227 07:54:41.820554 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b6af0c6_05e2_4adc_8e98_7d2e4f6254ea.slice/crio-af6923123932146675cbf7cab621b8d4c3d05f9d557e2883f54de4fb0c213f9f WatchSource:0}: Error finding container af6923123932146675cbf7cab621b8d4c3d05f9d557e2883f54de4fb0c213f9f: Status 404 returned error can't find the container with id af6923123932146675cbf7cab621b8d4c3d05f9d557e2883f54de4fb0c213f9f Dec 27 07:54:41 crc kubenswrapper[4934]: I1227 07:54:41.899662 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-79cf69ddc8-h47cv" event={"ID":"8b6af0c6-05e2-4adc-8e98-7d2e4f6254ea","Type":"ContainerStarted","Data":"af6923123932146675cbf7cab621b8d4c3d05f9d557e2883f54de4fb0c213f9f"} Dec 27 07:54:45 crc kubenswrapper[4934]: I1227 07:54:45.330253 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 07:54:45 crc kubenswrapper[4934]: I1227 07:54:45.330636 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 07:54:48 crc kubenswrapper[4934]: I1227 07:54:48.952379 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" event={"ID":"00799e47-9b94-49bf-8bd1-d6bb5036285a","Type":"ContainerStarted","Data":"ed64f69a04114620942ed0f14a24739b1d58dd40935c3a97801993860c86d312"} Dec 27 07:54:48 crc kubenswrapper[4934]: I1227 07:54:48.955116 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-79cf69ddc8-h47cv" event={"ID":"8b6af0c6-05e2-4adc-8e98-7d2e4f6254ea","Type":"ContainerStarted","Data":"e7ed0fca2335da4bd62b8d98ef8842fc297bb2e35929a624c730ca75dedec21d"} Dec 27 07:54:48 crc kubenswrapper[4934]: I1227 07:54:48.989958 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/cluster-logging-operator-79cf69ddc8-h47cv" podStartSLOduration=1.95135659 podStartE2EDuration="7.989937573s" podCreationTimestamp="2025-12-27 07:54:41 +0000 UTC" firstStartedPulling="2025-12-27 07:54:41.822513319 +0000 UTC m=+742.642953913" lastFinishedPulling="2025-12-27 07:54:47.861094302 +0000 UTC m=+748.681534896" observedRunningTime="2025-12-27 07:54:48.977956881 +0000 UTC m=+749.798397515" watchObservedRunningTime="2025-12-27 07:54:48.989937573 +0000 UTC m=+749.810378167" Dec 27 07:54:55 crc kubenswrapper[4934]: I1227 07:54:55.001483 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" event={"ID":"00799e47-9b94-49bf-8bd1-d6bb5036285a","Type":"ContainerStarted","Data":"4ba540919c08d4e7df62d23c8abdef8a9f7033370d3731fa794d0a65cadc7b38"} Dec 27 07:54:55 crc kubenswrapper[4934]: I1227 07:54:55.002364 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 07:54:55 crc kubenswrapper[4934]: I1227 07:54:55.006696 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 07:54:55 crc kubenswrapper[4934]: I1227 07:54:55.030390 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" podStartSLOduration=0.812440287 podStartE2EDuration="15.030367852s" podCreationTimestamp="2025-12-27 07:54:40 +0000 UTC" firstStartedPulling="2025-12-27 07:54:40.603822558 +0000 UTC m=+741.424263152" lastFinishedPulling="2025-12-27 07:54:54.821750123 +0000 UTC m=+755.642190717" observedRunningTime="2025-12-27 07:54:55.024257228 +0000 UTC m=+755.844697832" watchObservedRunningTime="2025-12-27 07:54:55.030367852 +0000 UTC m=+755.850808446" Dec 27 07:54:59 crc kubenswrapper[4934]: I1227 07:54:59.132759 4934 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 27 07:55:00 crc kubenswrapper[4934]: I1227 07:55:00.110173 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Dec 27 07:55:00 crc kubenswrapper[4934]: I1227 07:55:00.111165 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 27 07:55:00 crc kubenswrapper[4934]: I1227 07:55:00.115492 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Dec 27 07:55:00 crc kubenswrapper[4934]: I1227 07:55:00.121991 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 27 07:55:00 crc kubenswrapper[4934]: I1227 07:55:00.123935 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Dec 27 07:55:00 crc kubenswrapper[4934]: I1227 07:55:00.178619 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-16ad9f88-c1e4-49b7-b8a8-eec286508fae\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-16ad9f88-c1e4-49b7-b8a8-eec286508fae\") pod \"minio\" (UID: \"94eddaaa-f70e-4159-8f88-1874017e472c\") " pod="minio-dev/minio" Dec 27 07:55:00 crc kubenswrapper[4934]: I1227 07:55:00.178699 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgg26\" (UniqueName: \"kubernetes.io/projected/94eddaaa-f70e-4159-8f88-1874017e472c-kube-api-access-qgg26\") pod \"minio\" (UID: \"94eddaaa-f70e-4159-8f88-1874017e472c\") " pod="minio-dev/minio" Dec 27 07:55:00 crc kubenswrapper[4934]: I1227 07:55:00.280223 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-16ad9f88-c1e4-49b7-b8a8-eec286508fae\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-16ad9f88-c1e4-49b7-b8a8-eec286508fae\") pod \"minio\" (UID: \"94eddaaa-f70e-4159-8f88-1874017e472c\") " pod="minio-dev/minio" Dec 27 07:55:00 crc kubenswrapper[4934]: I1227 07:55:00.280370 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgg26\" (UniqueName: \"kubernetes.io/projected/94eddaaa-f70e-4159-8f88-1874017e472c-kube-api-access-qgg26\") pod \"minio\" (UID: \"94eddaaa-f70e-4159-8f88-1874017e472c\") " pod="minio-dev/minio" Dec 27 07:55:00 crc kubenswrapper[4934]: I1227 07:55:00.283921 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 07:55:00 crc kubenswrapper[4934]: I1227 07:55:00.283969 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-16ad9f88-c1e4-49b7-b8a8-eec286508fae\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-16ad9f88-c1e4-49b7-b8a8-eec286508fae\") pod \"minio\" (UID: \"94eddaaa-f70e-4159-8f88-1874017e472c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9cd05a04ed947d7896eee98f2ee8f96ea784f706610cadf01c76e2bff876ae6a/globalmount\"" pod="minio-dev/minio" Dec 27 07:55:00 crc kubenswrapper[4934]: I1227 07:55:00.303932 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgg26\" (UniqueName: \"kubernetes.io/projected/94eddaaa-f70e-4159-8f88-1874017e472c-kube-api-access-qgg26\") pod \"minio\" (UID: \"94eddaaa-f70e-4159-8f88-1874017e472c\") " pod="minio-dev/minio" Dec 27 07:55:00 crc kubenswrapper[4934]: I1227 07:55:00.309437 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-16ad9f88-c1e4-49b7-b8a8-eec286508fae\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-16ad9f88-c1e4-49b7-b8a8-eec286508fae\") pod \"minio\" (UID: \"94eddaaa-f70e-4159-8f88-1874017e472c\") " pod="minio-dev/minio" Dec 27 07:55:00 crc kubenswrapper[4934]: I1227 07:55:00.436450 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 27 07:55:00 crc kubenswrapper[4934]: I1227 07:55:00.942462 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 27 07:55:01 crc kubenswrapper[4934]: I1227 07:55:01.037607 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"94eddaaa-f70e-4159-8f88-1874017e472c","Type":"ContainerStarted","Data":"8e426f83aed6a211488fc29e7fb7fdeff645bd7bacbf0bb73bf5badc4d4cad11"} Dec 27 07:55:06 crc kubenswrapper[4934]: I1227 07:55:06.179244 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"94eddaaa-f70e-4159-8f88-1874017e472c","Type":"ContainerStarted","Data":"3e2edf464524900bc1f7df852600cfae3daacd21cb45781704bd273c5222ed5d"} Dec 27 07:55:06 crc kubenswrapper[4934]: I1227 07:55:06.207876 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=4.600112987 podStartE2EDuration="9.207860259s" podCreationTimestamp="2025-12-27 07:54:57 +0000 UTC" firstStartedPulling="2025-12-27 07:55:00.955679303 +0000 UTC m=+761.776119907" lastFinishedPulling="2025-12-27 07:55:05.563426585 +0000 UTC m=+766.383867179" observedRunningTime="2025-12-27 07:55:06.205174802 +0000 UTC m=+767.025615406" watchObservedRunningTime="2025-12-27 07:55:06.207860259 +0000 UTC m=+767.028300863" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.484398 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7"] Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.485625 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.492028 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-dockercfg-w8jjt" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.492275 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-grpc" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.493186 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-http" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.498516 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7"] Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.500707 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-config" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.500872 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-ca-bundle" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.579157 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5864b138-b8ef-405e-9ed9-be39f8e13e8d-config\") pod \"logging-loki-distributor-5f678c8dd6-nqqn7\" (UID: \"5864b138-b8ef-405e-9ed9-be39f8e13e8d\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.579239 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/5864b138-b8ef-405e-9ed9-be39f8e13e8d-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-5f678c8dd6-nqqn7\" (UID: \"5864b138-b8ef-405e-9ed9-be39f8e13e8d\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.579300 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/5864b138-b8ef-405e-9ed9-be39f8e13e8d-logging-loki-distributor-http\") pod \"logging-loki-distributor-5f678c8dd6-nqqn7\" (UID: \"5864b138-b8ef-405e-9ed9-be39f8e13e8d\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.579356 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nb4c\" (UniqueName: \"kubernetes.io/projected/5864b138-b8ef-405e-9ed9-be39f8e13e8d-kube-api-access-7nb4c\") pod \"logging-loki-distributor-5f678c8dd6-nqqn7\" (UID: \"5864b138-b8ef-405e-9ed9-be39f8e13e8d\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.579383 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5864b138-b8ef-405e-9ed9-be39f8e13e8d-logging-loki-ca-bundle\") pod \"logging-loki-distributor-5f678c8dd6-nqqn7\" (UID: \"5864b138-b8ef-405e-9ed9-be39f8e13e8d\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.646625 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-querier-76788598db-6wrwh"] Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.647864 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.651166 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-grpc" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.651562 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-s3" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.651733 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-http" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.659952 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-76788598db-6wrwh"] Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.681093 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5864b138-b8ef-405e-9ed9-be39f8e13e8d-config\") pod \"logging-loki-distributor-5f678c8dd6-nqqn7\" (UID: \"5864b138-b8ef-405e-9ed9-be39f8e13e8d\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.681151 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/5864b138-b8ef-405e-9ed9-be39f8e13e8d-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-5f678c8dd6-nqqn7\" (UID: \"5864b138-b8ef-405e-9ed9-be39f8e13e8d\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.681208 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/5864b138-b8ef-405e-9ed9-be39f8e13e8d-logging-loki-distributor-http\") pod \"logging-loki-distributor-5f678c8dd6-nqqn7\" (UID: \"5864b138-b8ef-405e-9ed9-be39f8e13e8d\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.681255 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nb4c\" (UniqueName: \"kubernetes.io/projected/5864b138-b8ef-405e-9ed9-be39f8e13e8d-kube-api-access-7nb4c\") pod \"logging-loki-distributor-5f678c8dd6-nqqn7\" (UID: \"5864b138-b8ef-405e-9ed9-be39f8e13e8d\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.681278 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5864b138-b8ef-405e-9ed9-be39f8e13e8d-logging-loki-ca-bundle\") pod \"logging-loki-distributor-5f678c8dd6-nqqn7\" (UID: \"5864b138-b8ef-405e-9ed9-be39f8e13e8d\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.682027 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5864b138-b8ef-405e-9ed9-be39f8e13e8d-logging-loki-ca-bundle\") pod \"logging-loki-distributor-5f678c8dd6-nqqn7\" (UID: \"5864b138-b8ef-405e-9ed9-be39f8e13e8d\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.682793 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5864b138-b8ef-405e-9ed9-be39f8e13e8d-config\") pod \"logging-loki-distributor-5f678c8dd6-nqqn7\" (UID: \"5864b138-b8ef-405e-9ed9-be39f8e13e8d\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.691942 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/5864b138-b8ef-405e-9ed9-be39f8e13e8d-logging-loki-distributor-http\") pod \"logging-loki-distributor-5f678c8dd6-nqqn7\" (UID: \"5864b138-b8ef-405e-9ed9-be39f8e13e8d\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.691952 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/5864b138-b8ef-405e-9ed9-be39f8e13e8d-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-5f678c8dd6-nqqn7\" (UID: \"5864b138-b8ef-405e-9ed9-be39f8e13e8d\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.713937 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nb4c\" (UniqueName: \"kubernetes.io/projected/5864b138-b8ef-405e-9ed9-be39f8e13e8d-kube-api-access-7nb4c\") pod \"logging-loki-distributor-5f678c8dd6-nqqn7\" (UID: \"5864b138-b8ef-405e-9ed9-be39f8e13e8d\") " pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.745176 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt"] Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.746153 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.748288 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-grpc" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.748616 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-http" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.760398 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt"] Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.783347 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae25e4b9-7a2a-41ff-9944-d6c45603cbec-config\") pod \"logging-loki-querier-76788598db-6wrwh\" (UID: \"ae25e4b9-7a2a-41ff-9944-d6c45603cbec\") " pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.783422 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/ae25e4b9-7a2a-41ff-9944-d6c45603cbec-logging-loki-querier-grpc\") pod \"logging-loki-querier-76788598db-6wrwh\" (UID: \"ae25e4b9-7a2a-41ff-9944-d6c45603cbec\") " pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.783465 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/ae25e4b9-7a2a-41ff-9944-d6c45603cbec-logging-loki-s3\") pod \"logging-loki-querier-76788598db-6wrwh\" (UID: \"ae25e4b9-7a2a-41ff-9944-d6c45603cbec\") " pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.783487 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/ae25e4b9-7a2a-41ff-9944-d6c45603cbec-logging-loki-querier-http\") pod \"logging-loki-querier-76788598db-6wrwh\" (UID: \"ae25e4b9-7a2a-41ff-9944-d6c45603cbec\") " pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.783501 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae25e4b9-7a2a-41ff-9944-d6c45603cbec-logging-loki-ca-bundle\") pod \"logging-loki-querier-76788598db-6wrwh\" (UID: \"ae25e4b9-7a2a-41ff-9944-d6c45603cbec\") " pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.783529 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88zcg\" (UniqueName: \"kubernetes.io/projected/ae25e4b9-7a2a-41ff-9944-d6c45603cbec-kube-api-access-88zcg\") pod \"logging-loki-querier-76788598db-6wrwh\" (UID: \"ae25e4b9-7a2a-41ff-9944-d6c45603cbec\") " pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.804749 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.842242 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-74dbf44c45-qd69g"] Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.852747 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.857183 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-http" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.876649 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.876775 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.876657 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway-ca-bundle" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.876938 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-client-http" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.876913 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-dockercfg-7t7f7" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.878769 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v"] Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.924642 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.929436 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/47cbee2e-c325-4a8d-a26d-94bced60ee78-tenants\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.929507 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47cbee2e-c325-4a8d-a26d-94bced60ee78-logging-loki-ca-bundle\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.929544 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae25e4b9-7a2a-41ff-9944-d6c45603cbec-config\") pod \"logging-loki-querier-76788598db-6wrwh\" (UID: \"ae25e4b9-7a2a-41ff-9944-d6c45603cbec\") " pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.929570 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9e3cc00-7197-4c75-a732-136e65d893c8-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-69d9546745-4b9rt\" (UID: \"b9e3cc00-7197-4c75-a732-136e65d893c8\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.929782 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9e3cc00-7197-4c75-a732-136e65d893c8-config\") pod \"logging-loki-query-frontend-69d9546745-4b9rt\" (UID: \"b9e3cc00-7197-4c75-a732-136e65d893c8\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.929817 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/b9e3cc00-7197-4c75-a732-136e65d893c8-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-69d9546745-4b9rt\" (UID: \"b9e3cc00-7197-4c75-a732-136e65d893c8\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.929843 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75ckr\" (UniqueName: \"kubernetes.io/projected/b9e3cc00-7197-4c75-a732-136e65d893c8-kube-api-access-75ckr\") pod \"logging-loki-query-frontend-69d9546745-4b9rt\" (UID: \"b9e3cc00-7197-4c75-a732-136e65d893c8\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.929872 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47cbee2e-c325-4a8d-a26d-94bced60ee78-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.929896 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/ae25e4b9-7a2a-41ff-9944-d6c45603cbec-logging-loki-querier-grpc\") pod \"logging-loki-querier-76788598db-6wrwh\" (UID: \"ae25e4b9-7a2a-41ff-9944-d6c45603cbec\") " pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.929915 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/b9e3cc00-7197-4c75-a732-136e65d893c8-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-69d9546745-4b9rt\" (UID: \"b9e3cc00-7197-4c75-a732-136e65d893c8\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.929939 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/47cbee2e-c325-4a8d-a26d-94bced60ee78-rbac\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.929960 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/ae25e4b9-7a2a-41ff-9944-d6c45603cbec-logging-loki-s3\") pod \"logging-loki-querier-76788598db-6wrwh\" (UID: \"ae25e4b9-7a2a-41ff-9944-d6c45603cbec\") " pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.929982 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/47cbee2e-c325-4a8d-a26d-94bced60ee78-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.930008 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/ae25e4b9-7a2a-41ff-9944-d6c45603cbec-logging-loki-querier-http\") pod \"logging-loki-querier-76788598db-6wrwh\" (UID: \"ae25e4b9-7a2a-41ff-9944-d6c45603cbec\") " pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.930028 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae25e4b9-7a2a-41ff-9944-d6c45603cbec-logging-loki-ca-bundle\") pod \"logging-loki-querier-76788598db-6wrwh\" (UID: \"ae25e4b9-7a2a-41ff-9944-d6c45603cbec\") " pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.930051 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88ndt\" (UniqueName: \"kubernetes.io/projected/47cbee2e-c325-4a8d-a26d-94bced60ee78-kube-api-access-88ndt\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.930074 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88zcg\" (UniqueName: \"kubernetes.io/projected/ae25e4b9-7a2a-41ff-9944-d6c45603cbec-kube-api-access-88zcg\") pod \"logging-loki-querier-76788598db-6wrwh\" (UID: \"ae25e4b9-7a2a-41ff-9944-d6c45603cbec\") " pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.930133 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/47cbee2e-c325-4a8d-a26d-94bced60ee78-lokistack-gateway\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.930163 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/47cbee2e-c325-4a8d-a26d-94bced60ee78-tls-secret\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.931162 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae25e4b9-7a2a-41ff-9944-d6c45603cbec-config\") pod \"logging-loki-querier-76788598db-6wrwh\" (UID: \"ae25e4b9-7a2a-41ff-9944-d6c45603cbec\") " pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.947752 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae25e4b9-7a2a-41ff-9944-d6c45603cbec-logging-loki-ca-bundle\") pod \"logging-loki-querier-76788598db-6wrwh\" (UID: \"ae25e4b9-7a2a-41ff-9944-d6c45603cbec\") " pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.957531 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/ae25e4b9-7a2a-41ff-9944-d6c45603cbec-logging-loki-s3\") pod \"logging-loki-querier-76788598db-6wrwh\" (UID: \"ae25e4b9-7a2a-41ff-9944-d6c45603cbec\") " pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.958473 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-74dbf44c45-qd69g"] Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.969177 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/ae25e4b9-7a2a-41ff-9944-d6c45603cbec-logging-loki-querier-grpc\") pod \"logging-loki-querier-76788598db-6wrwh\" (UID: \"ae25e4b9-7a2a-41ff-9944-d6c45603cbec\") " pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.978844 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/ae25e4b9-7a2a-41ff-9944-d6c45603cbec-logging-loki-querier-http\") pod \"logging-loki-querier-76788598db-6wrwh\" (UID: \"ae25e4b9-7a2a-41ff-9944-d6c45603cbec\") " pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:10 crc kubenswrapper[4934]: I1227 07:55:10.996143 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v"] Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.000599 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88zcg\" (UniqueName: \"kubernetes.io/projected/ae25e4b9-7a2a-41ff-9944-d6c45603cbec-kube-api-access-88zcg\") pod \"logging-loki-querier-76788598db-6wrwh\" (UID: \"ae25e4b9-7a2a-41ff-9944-d6c45603cbec\") " pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.032331 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eac5b26c-d598-4c68-95b5-e2583b456af9-logging-loki-ca-bundle\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.032375 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/47cbee2e-c325-4a8d-a26d-94bced60ee78-tls-secret\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.032410 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbf5f\" (UniqueName: \"kubernetes.io/projected/eac5b26c-d598-4c68-95b5-e2583b456af9-kube-api-access-wbf5f\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.032432 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/47cbee2e-c325-4a8d-a26d-94bced60ee78-tenants\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.032446 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/eac5b26c-d598-4c68-95b5-e2583b456af9-rbac\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.032469 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47cbee2e-c325-4a8d-a26d-94bced60ee78-logging-loki-ca-bundle\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.032488 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9e3cc00-7197-4c75-a732-136e65d893c8-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-69d9546745-4b9rt\" (UID: \"b9e3cc00-7197-4c75-a732-136e65d893c8\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.032503 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9e3cc00-7197-4c75-a732-136e65d893c8-config\") pod \"logging-loki-query-frontend-69d9546745-4b9rt\" (UID: \"b9e3cc00-7197-4c75-a732-136e65d893c8\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.032526 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/b9e3cc00-7197-4c75-a732-136e65d893c8-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-69d9546745-4b9rt\" (UID: \"b9e3cc00-7197-4c75-a732-136e65d893c8\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.032542 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/eac5b26c-d598-4c68-95b5-e2583b456af9-tenants\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.032561 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75ckr\" (UniqueName: \"kubernetes.io/projected/b9e3cc00-7197-4c75-a732-136e65d893c8-kube-api-access-75ckr\") pod \"logging-loki-query-frontend-69d9546745-4b9rt\" (UID: \"b9e3cc00-7197-4c75-a732-136e65d893c8\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.032582 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/eac5b26c-d598-4c68-95b5-e2583b456af9-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.032598 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eac5b26c-d598-4c68-95b5-e2583b456af9-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.032619 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47cbee2e-c325-4a8d-a26d-94bced60ee78-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.032641 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/b9e3cc00-7197-4c75-a732-136e65d893c8-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-69d9546745-4b9rt\" (UID: \"b9e3cc00-7197-4c75-a732-136e65d893c8\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.032661 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/47cbee2e-c325-4a8d-a26d-94bced60ee78-rbac\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.032679 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/47cbee2e-c325-4a8d-a26d-94bced60ee78-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.032697 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/eac5b26c-d598-4c68-95b5-e2583b456af9-tls-secret\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.032718 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88ndt\" (UniqueName: \"kubernetes.io/projected/47cbee2e-c325-4a8d-a26d-94bced60ee78-kube-api-access-88ndt\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.032734 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/eac5b26c-d598-4c68-95b5-e2583b456af9-lokistack-gateway\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.032755 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/47cbee2e-c325-4a8d-a26d-94bced60ee78-lokistack-gateway\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.033491 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/47cbee2e-c325-4a8d-a26d-94bced60ee78-lokistack-gateway\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:11 crc kubenswrapper[4934]: E1227 07:55:11.033564 4934 secret.go:188] Couldn't get secret openshift-logging/logging-loki-gateway-http: secret "logging-loki-gateway-http" not found Dec 27 07:55:11 crc kubenswrapper[4934]: E1227 07:55:11.033603 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/47cbee2e-c325-4a8d-a26d-94bced60ee78-tls-secret podName:47cbee2e-c325-4a8d-a26d-94bced60ee78 nodeName:}" failed. No retries permitted until 2025-12-27 07:55:11.533590215 +0000 UTC m=+772.354030809 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/47cbee2e-c325-4a8d-a26d-94bced60ee78-tls-secret") pod "logging-loki-gateway-74dbf44c45-qd69g" (UID: "47cbee2e-c325-4a8d-a26d-94bced60ee78") : secret "logging-loki-gateway-http" not found Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.034643 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47cbee2e-c325-4a8d-a26d-94bced60ee78-logging-loki-ca-bundle\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.035205 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b9e3cc00-7197-4c75-a732-136e65d893c8-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-69d9546745-4b9rt\" (UID: \"b9e3cc00-7197-4c75-a732-136e65d893c8\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.035795 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9e3cc00-7197-4c75-a732-136e65d893c8-config\") pod \"logging-loki-query-frontend-69d9546745-4b9rt\" (UID: \"b9e3cc00-7197-4c75-a732-136e65d893c8\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.041660 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/47cbee2e-c325-4a8d-a26d-94bced60ee78-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.042976 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/47cbee2e-c325-4a8d-a26d-94bced60ee78-rbac\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.056826 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/47cbee2e-c325-4a8d-a26d-94bced60ee78-tenants\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.058531 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/47cbee2e-c325-4a8d-a26d-94bced60ee78-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.058978 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/b9e3cc00-7197-4c75-a732-136e65d893c8-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-69d9546745-4b9rt\" (UID: \"b9e3cc00-7197-4c75-a732-136e65d893c8\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.064997 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88ndt\" (UniqueName: \"kubernetes.io/projected/47cbee2e-c325-4a8d-a26d-94bced60ee78-kube-api-access-88ndt\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.065785 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75ckr\" (UniqueName: \"kubernetes.io/projected/b9e3cc00-7197-4c75-a732-136e65d893c8-kube-api-access-75ckr\") pod \"logging-loki-query-frontend-69d9546745-4b9rt\" (UID: \"b9e3cc00-7197-4c75-a732-136e65d893c8\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.066379 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/b9e3cc00-7197-4c75-a732-136e65d893c8-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-69d9546745-4b9rt\" (UID: \"b9e3cc00-7197-4c75-a732-136e65d893c8\") " pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.066786 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.137791 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/eac5b26c-d598-4c68-95b5-e2583b456af9-tenants\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.138161 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/eac5b26c-d598-4c68-95b5-e2583b456af9-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.138187 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eac5b26c-d598-4c68-95b5-e2583b456af9-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.138227 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/eac5b26c-d598-4c68-95b5-e2583b456af9-tls-secret\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.138254 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/eac5b26c-d598-4c68-95b5-e2583b456af9-lokistack-gateway\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.138282 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eac5b26c-d598-4c68-95b5-e2583b456af9-logging-loki-ca-bundle\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.138327 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbf5f\" (UniqueName: \"kubernetes.io/projected/eac5b26c-d598-4c68-95b5-e2583b456af9-kube-api-access-wbf5f\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.138347 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/eac5b26c-d598-4c68-95b5-e2583b456af9-rbac\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: E1227 07:55:11.138390 4934 secret.go:188] Couldn't get secret openshift-logging/logging-loki-gateway-http: secret "logging-loki-gateway-http" not found Dec 27 07:55:11 crc kubenswrapper[4934]: E1227 07:55:11.138488 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eac5b26c-d598-4c68-95b5-e2583b456af9-tls-secret podName:eac5b26c-d598-4c68-95b5-e2583b456af9 nodeName:}" failed. No retries permitted until 2025-12-27 07:55:11.638470294 +0000 UTC m=+772.458910888 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/eac5b26c-d598-4c68-95b5-e2583b456af9-tls-secret") pod "logging-loki-gateway-74dbf44c45-4ll7v" (UID: "eac5b26c-d598-4c68-95b5-e2583b456af9") : secret "logging-loki-gateway-http" not found Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.139198 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/eac5b26c-d598-4c68-95b5-e2583b456af9-rbac\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.139775 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/eac5b26c-d598-4c68-95b5-e2583b456af9-lokistack-gateway\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.140343 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eac5b26c-d598-4c68-95b5-e2583b456af9-logging-loki-ca-bundle\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.140785 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/eac5b26c-d598-4c68-95b5-e2583b456af9-tenants\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.141168 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eac5b26c-d598-4c68-95b5-e2583b456af9-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.142757 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/eac5b26c-d598-4c68-95b5-e2583b456af9-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.161884 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbf5f\" (UniqueName: \"kubernetes.io/projected/eac5b26c-d598-4c68-95b5-e2583b456af9-kube-api-access-wbf5f\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.266206 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.362985 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt"] Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.411446 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7"] Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.548464 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/47cbee2e-c325-4a8d-a26d-94bced60ee78-tls-secret\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.551697 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/47cbee2e-c325-4a8d-a26d-94bced60ee78-tls-secret\") pod \"logging-loki-gateway-74dbf44c45-qd69g\" (UID: \"47cbee2e-c325-4a8d-a26d-94bced60ee78\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.558514 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.649630 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/eac5b26c-d598-4c68-95b5-e2583b456af9-tls-secret\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.650192 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.651044 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.654899 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-grpc" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.655489 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/eac5b26c-d598-4c68-95b5-e2583b456af9-tls-secret\") pod \"logging-loki-gateway-74dbf44c45-4ll7v\" (UID: \"eac5b26c-d598-4c68-95b5-e2583b456af9\") " pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.658307 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-http" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.661619 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.704912 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-76788598db-6wrwh"] Dec 27 07:55:11 crc kubenswrapper[4934]: W1227 07:55:11.709876 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae25e4b9_7a2a_41ff_9944_d6c45603cbec.slice/crio-29f0e003a2ce9333530b577fc6472a0ae451a98517031379367ffc8a6412c8a1 WatchSource:0}: Error finding container 29f0e003a2ce9333530b577fc6472a0ae451a98517031379367ffc8a6412c8a1: Status 404 returned error can't find the container with id 29f0e003a2ce9333530b577fc6472a0ae451a98517031379367ffc8a6412c8a1 Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.731303 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.732235 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.735152 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-http" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.735457 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-grpc" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.738939 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.751007 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/642f3213-8063-4ade-800e-4e05ba16d642-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.751053 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ea8f1f02-868e-4794-9ae2-03d526a39bfb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ea8f1f02-868e-4794-9ae2-03d526a39bfb\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.751099 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvp7s\" (UniqueName: \"kubernetes.io/projected/642f3213-8063-4ade-800e-4e05ba16d642-kube-api-access-nvp7s\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.751136 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/642f3213-8063-4ade-800e-4e05ba16d642-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.751151 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/642f3213-8063-4ade-800e-4e05ba16d642-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.751189 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/642f3213-8063-4ade-800e-4e05ba16d642-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.751227 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/642f3213-8063-4ade-800e-4e05ba16d642-config\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.751252 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-4b09b737-217a-4d43-8c05-16ea8905812d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b09b737-217a-4d43-8c05-16ea8905812d\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.814307 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.815152 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.817022 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-grpc" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.817049 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-http" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.822970 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.852724 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6sgz\" (UniqueName: \"kubernetes.io/projected/e6de9198-9266-43e8-b760-e21f0afa885e-kube-api-access-k6sgz\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") " pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.852797 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/e6de9198-9266-43e8-b760-e21f0afa885e-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") " pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.852841 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/642f3213-8063-4ade-800e-4e05ba16d642-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.852871 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/01f91788-9f82-4508-906b-1e98c9e05c2c-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.852907 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmmhl\" (UniqueName: \"kubernetes.io/projected/01f91788-9f82-4508-906b-1e98c9e05c2c-kube-api-access-jmmhl\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.852940 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/642f3213-8063-4ade-800e-4e05ba16d642-config\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.852975 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/01f91788-9f82-4508-906b-1e98c9e05c2c-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.853010 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01f91788-9f82-4508-906b-1e98c9e05c2c-config\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.853048 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6de9198-9266-43e8-b760-e21f0afa885e-config\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") " pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.853104 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/642f3213-8063-4ade-800e-4e05ba16d642-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.853151 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-baf2d1d0-3246-4576-b1ea-1bfb9b2c2de5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-baf2d1d0-3246-4576-b1ea-1bfb9b2c2de5\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") " pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.853196 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/e6de9198-9266-43e8-b760-e21f0afa885e-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") " pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.853234 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/e6de9198-9266-43e8-b760-e21f0afa885e-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") " pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.853271 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/642f3213-8063-4ade-800e-4e05ba16d642-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.853300 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/642f3213-8063-4ade-800e-4e05ba16d642-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.853344 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/01f91788-9f82-4508-906b-1e98c9e05c2c-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.853380 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b0cca3fc-49f4-40e4-bf3f-5ac85023e2bc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b0cca3fc-49f4-40e4-bf3f-5ac85023e2bc\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.853421 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-4b09b737-217a-4d43-8c05-16ea8905812d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b09b737-217a-4d43-8c05-16ea8905812d\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.853461 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6de9198-9266-43e8-b760-e21f0afa885e-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") " pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.853514 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ea8f1f02-868e-4794-9ae2-03d526a39bfb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ea8f1f02-868e-4794-9ae2-03d526a39bfb\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.853555 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvp7s\" (UniqueName: \"kubernetes.io/projected/642f3213-8063-4ade-800e-4e05ba16d642-kube-api-access-nvp7s\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.853604 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/01f91788-9f82-4508-906b-1e98c9e05c2c-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.854927 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/642f3213-8063-4ade-800e-4e05ba16d642-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.855131 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/642f3213-8063-4ade-800e-4e05ba16d642-config\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.857348 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/642f3213-8063-4ade-800e-4e05ba16d642-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.857814 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.857853 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-4b09b737-217a-4d43-8c05-16ea8905812d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b09b737-217a-4d43-8c05-16ea8905812d\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/6bbb70cac5ab7dad7afbee945be5bf81cae29e64f5a6410165f4656ca2ff47f2/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.857917 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.857942 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ea8f1f02-868e-4794-9ae2-03d526a39bfb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ea8f1f02-868e-4794-9ae2-03d526a39bfb\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/fad2b680c383300c8de382ba71ef27a306ea92766e661a5df5b6af5073ac1c6a/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.859498 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/642f3213-8063-4ade-800e-4e05ba16d642-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.863735 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/642f3213-8063-4ade-800e-4e05ba16d642-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.872502 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvp7s\" (UniqueName: \"kubernetes.io/projected/642f3213-8063-4ade-800e-4e05ba16d642-kube-api-access-nvp7s\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.883417 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-4b09b737-217a-4d43-8c05-16ea8905812d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-4b09b737-217a-4d43-8c05-16ea8905812d\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.889775 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ea8f1f02-868e-4794-9ae2-03d526a39bfb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ea8f1f02-868e-4794-9ae2-03d526a39bfb\") pod \"logging-loki-ingester-0\" (UID: \"642f3213-8063-4ade-800e-4e05ba16d642\") " pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.932096 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.955494 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/e6de9198-9266-43e8-b760-e21f0afa885e-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") " pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.955675 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/01f91788-9f82-4508-906b-1e98c9e05c2c-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.955750 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmmhl\" (UniqueName: \"kubernetes.io/projected/01f91788-9f82-4508-906b-1e98c9e05c2c-kube-api-access-jmmhl\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.955804 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/01f91788-9f82-4508-906b-1e98c9e05c2c-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.955891 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01f91788-9f82-4508-906b-1e98c9e05c2c-config\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.956194 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6de9198-9266-43e8-b760-e21f0afa885e-config\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") " pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.956282 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-baf2d1d0-3246-4576-b1ea-1bfb9b2c2de5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-baf2d1d0-3246-4576-b1ea-1bfb9b2c2de5\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") " pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.956335 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/e6de9198-9266-43e8-b760-e21f0afa885e-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") " pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.956379 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/e6de9198-9266-43e8-b760-e21f0afa885e-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") " pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.956425 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/01f91788-9f82-4508-906b-1e98c9e05c2c-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.956474 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b0cca3fc-49f4-40e4-bf3f-5ac85023e2bc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b0cca3fc-49f4-40e4-bf3f-5ac85023e2bc\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.956550 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6de9198-9266-43e8-b760-e21f0afa885e-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") " pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.956625 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/01f91788-9f82-4508-906b-1e98c9e05c2c-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.956667 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6sgz\" (UniqueName: \"kubernetes.io/projected/e6de9198-9266-43e8-b760-e21f0afa885e-kube-api-access-k6sgz\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") " pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.957419 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6de9198-9266-43e8-b760-e21f0afa885e-config\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") " pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.957618 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01f91788-9f82-4508-906b-1e98c9e05c2c-config\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.958212 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6de9198-9266-43e8-b760-e21f0afa885e-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") " pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.958275 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/01f91788-9f82-4508-906b-1e98c9e05c2c-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.960265 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/e6de9198-9266-43e8-b760-e21f0afa885e-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") " pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.961227 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.961287 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b0cca3fc-49f4-40e4-bf3f-5ac85023e2bc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b0cca3fc-49f4-40e4-bf3f-5ac85023e2bc\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/27b9261bf19104d461dd7ea5d0c12b6628d0d3f33de08cfa60ca3602dc2cd2e2/globalmount\"" pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.963476 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/01f91788-9f82-4508-906b-1e98c9e05c2c-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.967714 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/01f91788-9f82-4508-906b-1e98c9e05c2c-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.967968 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/01f91788-9f82-4508-906b-1e98c9e05c2c-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.968947 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.968984 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-baf2d1d0-3246-4576-b1ea-1bfb9b2c2de5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-baf2d1d0-3246-4576-b1ea-1bfb9b2c2de5\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5e6ab52ea6e516c24f08d2a7526d498eed2eefd0b6a686a97305681d5d7af397/globalmount\"" pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.969171 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/e6de9198-9266-43e8-b760-e21f0afa885e-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") " pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.969815 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/e6de9198-9266-43e8-b760-e21f0afa885e-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") " pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.974495 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6sgz\" (UniqueName: \"kubernetes.io/projected/e6de9198-9266-43e8-b760-e21f0afa885e-kube-api-access-k6sgz\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") " pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.979233 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmmhl\" (UniqueName: \"kubernetes.io/projected/01f91788-9f82-4508-906b-1e98c9e05c2c-kube-api-access-jmmhl\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.980281 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.990783 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b0cca3fc-49f4-40e4-bf3f-5ac85023e2bc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b0cca3fc-49f4-40e4-bf3f-5ac85023e2bc\") pod \"logging-loki-index-gateway-0\" (UID: \"01f91788-9f82-4508-906b-1e98c9e05c2c\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:11 crc kubenswrapper[4934]: I1227 07:55:11.994450 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-baf2d1d0-3246-4576-b1ea-1bfb9b2c2de5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-baf2d1d0-3246-4576-b1ea-1bfb9b2c2de5\") pod \"logging-loki-compactor-0\" (UID: \"e6de9198-9266-43e8-b760-e21f0afa885e\") " pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:12 crc kubenswrapper[4934]: I1227 07:55:12.005585 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-74dbf44c45-qd69g"] Dec 27 07:55:12 crc kubenswrapper[4934]: I1227 07:55:12.064710 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:12 crc kubenswrapper[4934]: I1227 07:55:12.132318 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:12 crc kubenswrapper[4934]: I1227 07:55:12.212610 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 27 07:55:12 crc kubenswrapper[4934]: W1227 07:55:12.228266 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod642f3213_8063_4ade_800e_4e05ba16d642.slice/crio-ecb13fbce4a8e378861ab3670db53cf7d77ee5d479219f0396ef4789c7b166f8 WatchSource:0}: Error finding container ecb13fbce4a8e378861ab3670db53cf7d77ee5d479219f0396ef4789c7b166f8: Status 404 returned error can't find the container with id ecb13fbce4a8e378861ab3670db53cf7d77ee5d479219f0396ef4789c7b166f8 Dec 27 07:55:12 crc kubenswrapper[4934]: I1227 07:55:12.228625 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" event={"ID":"47cbee2e-c325-4a8d-a26d-94bced60ee78","Type":"ContainerStarted","Data":"27c43af5f26d604397ce1ffd54c4127b02ed3be9a498432f29b00dcec40477b3"} Dec 27 07:55:12 crc kubenswrapper[4934]: I1227 07:55:12.229710 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" event={"ID":"ae25e4b9-7a2a-41ff-9944-d6c45603cbec","Type":"ContainerStarted","Data":"29f0e003a2ce9333530b577fc6472a0ae451a98517031379367ffc8a6412c8a1"} Dec 27 07:55:12 crc kubenswrapper[4934]: I1227 07:55:12.231688 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" event={"ID":"b9e3cc00-7197-4c75-a732-136e65d893c8","Type":"ContainerStarted","Data":"0393b6ff7252ea5dd80b86c5a0100d08ad7a48799a6598da1d739cf2da3982e1"} Dec 27 07:55:12 crc kubenswrapper[4934]: I1227 07:55:12.234521 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" event={"ID":"5864b138-b8ef-405e-9ed9-be39f8e13e8d","Type":"ContainerStarted","Data":"acfad1010f5818bef9955fdbe3d2185a1200115fea284adfbfd9feb63aefa30d"} Dec 27 07:55:12 crc kubenswrapper[4934]: I1227 07:55:12.350480 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v"] Dec 27 07:55:12 crc kubenswrapper[4934]: W1227 07:55:12.353264 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeac5b26c_d598_4c68_95b5_e2583b456af9.slice/crio-b0443f42c8e15cf0f327c5dddf181cc3bca05623962b8f4ccd689f21e1ba2f8c WatchSource:0}: Error finding container b0443f42c8e15cf0f327c5dddf181cc3bca05623962b8f4ccd689f21e1ba2f8c: Status 404 returned error can't find the container with id b0443f42c8e15cf0f327c5dddf181cc3bca05623962b8f4ccd689f21e1ba2f8c Dec 27 07:55:12 crc kubenswrapper[4934]: I1227 07:55:12.518268 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 27 07:55:12 crc kubenswrapper[4934]: W1227 07:55:12.521554 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6de9198_9266_43e8_b760_e21f0afa885e.slice/crio-abf9edf91e45fd94f1173649560e87f2b74fb2b688aff87f9a06472f161380fb WatchSource:0}: Error finding container abf9edf91e45fd94f1173649560e87f2b74fb2b688aff87f9a06472f161380fb: Status 404 returned error can't find the container with id abf9edf91e45fd94f1173649560e87f2b74fb2b688aff87f9a06472f161380fb Dec 27 07:55:12 crc kubenswrapper[4934]: I1227 07:55:12.579892 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 27 07:55:12 crc kubenswrapper[4934]: W1227 07:55:12.592858 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01f91788_9f82_4508_906b_1e98c9e05c2c.slice/crio-70ccaeef12e7b1d808fe4d719214450a21a3b3bb68a56f29be6e7285245a0293 WatchSource:0}: Error finding container 70ccaeef12e7b1d808fe4d719214450a21a3b3bb68a56f29be6e7285245a0293: Status 404 returned error can't find the container with id 70ccaeef12e7b1d808fe4d719214450a21a3b3bb68a56f29be6e7285245a0293 Dec 27 07:55:13 crc kubenswrapper[4934]: I1227 07:55:13.241487 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"01f91788-9f82-4508-906b-1e98c9e05c2c","Type":"ContainerStarted","Data":"70ccaeef12e7b1d808fe4d719214450a21a3b3bb68a56f29be6e7285245a0293"} Dec 27 07:55:13 crc kubenswrapper[4934]: I1227 07:55:13.243000 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"642f3213-8063-4ade-800e-4e05ba16d642","Type":"ContainerStarted","Data":"ecb13fbce4a8e378861ab3670db53cf7d77ee5d479219f0396ef4789c7b166f8"} Dec 27 07:55:13 crc kubenswrapper[4934]: I1227 07:55:13.244390 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"e6de9198-9266-43e8-b760-e21f0afa885e","Type":"ContainerStarted","Data":"abf9edf91e45fd94f1173649560e87f2b74fb2b688aff87f9a06472f161380fb"} Dec 27 07:55:13 crc kubenswrapper[4934]: I1227 07:55:13.245330 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" event={"ID":"eac5b26c-d598-4c68-95b5-e2583b456af9","Type":"ContainerStarted","Data":"b0443f42c8e15cf0f327c5dddf181cc3bca05623962b8f4ccd689f21e1ba2f8c"} Dec 27 07:55:15 crc kubenswrapper[4934]: I1227 07:55:15.330667 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 07:55:15 crc kubenswrapper[4934]: I1227 07:55:15.330742 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 07:55:15 crc kubenswrapper[4934]: I1227 07:55:15.330791 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 07:55:15 crc kubenswrapper[4934]: I1227 07:55:15.331522 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"56bf8d198d1688bbc71036fbf874230b44fa217f3818e5e1d2117e459b2b5477"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 07:55:15 crc kubenswrapper[4934]: I1227 07:55:15.331587 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://56bf8d198d1688bbc71036fbf874230b44fa217f3818e5e1d2117e459b2b5477" gracePeriod=600 Dec 27 07:55:16 crc kubenswrapper[4934]: I1227 07:55:16.266963 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="56bf8d198d1688bbc71036fbf874230b44fa217f3818e5e1d2117e459b2b5477" exitCode=0 Dec 27 07:55:16 crc kubenswrapper[4934]: I1227 07:55:16.267034 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"56bf8d198d1688bbc71036fbf874230b44fa217f3818e5e1d2117e459b2b5477"} Dec 27 07:55:16 crc kubenswrapper[4934]: I1227 07:55:16.267432 4934 scope.go:117] "RemoveContainer" containerID="d237f7eb348f6dbacefb4e9dfc24791b0b82ed817b960e703b17f6f47221575e" Dec 27 07:55:17 crc kubenswrapper[4934]: I1227 07:55:17.274820 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" event={"ID":"5864b138-b8ef-405e-9ed9-be39f8e13e8d","Type":"ContainerStarted","Data":"ac59837da57a6f1b3b1202f2113f86de18e46324a1c1e90d3b6a627c6a5acc62"} Dec 27 07:55:17 crc kubenswrapper[4934]: I1227 07:55:17.275436 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" Dec 27 07:55:17 crc kubenswrapper[4934]: I1227 07:55:17.276277 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"e6de9198-9266-43e8-b760-e21f0afa885e","Type":"ContainerStarted","Data":"a9259fb1bbe4a549f45ad0eb499811b8a00831562110671d5a7af5551a94d197"} Dec 27 07:55:17 crc kubenswrapper[4934]: I1227 07:55:17.276441 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:17 crc kubenswrapper[4934]: I1227 07:55:17.278441 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" event={"ID":"b9e3cc00-7197-4c75-a732-136e65d893c8","Type":"ContainerStarted","Data":"b373dd36cc9418a8b7c7fa71e36544f25278ab5b43523449edee67bfa5cd6e6c"} Dec 27 07:55:17 crc kubenswrapper[4934]: I1227 07:55:17.278612 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" Dec 27 07:55:17 crc kubenswrapper[4934]: I1227 07:55:17.281063 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"4ea115bb4d50c7db66f5a9c8846c893a8cb9f55d01ba76e6418fe5095edfd343"} Dec 27 07:55:17 crc kubenswrapper[4934]: I1227 07:55:17.283312 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"642f3213-8063-4ade-800e-4e05ba16d642","Type":"ContainerStarted","Data":"841699254c8f6108750191034a0cd5dedd4f1fe5b6e79862bdff9790b048bc46"} Dec 27 07:55:17 crc kubenswrapper[4934]: I1227 07:55:17.283396 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:55:17 crc kubenswrapper[4934]: I1227 07:55:17.285582 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" event={"ID":"47cbee2e-c325-4a8d-a26d-94bced60ee78","Type":"ContainerStarted","Data":"e3c22b0d1ae4de8b1806ada511a18659b2c8be444b5f79d653030d33fe15b701"} Dec 27 07:55:17 crc kubenswrapper[4934]: I1227 07:55:17.287551 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" event={"ID":"eac5b26c-d598-4c68-95b5-e2583b456af9","Type":"ContainerStarted","Data":"a2ccae647164c50fdd1e5bb749d92cd83264e96845f70b4a19b30d702dfcd46d"} Dec 27 07:55:17 crc kubenswrapper[4934]: I1227 07:55:17.288974 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" event={"ID":"ae25e4b9-7a2a-41ff-9944-d6c45603cbec","Type":"ContainerStarted","Data":"b3c160759a9272165da4179575f11ec97b7fe5905ef590604a9414a882ef1505"} Dec 27 07:55:17 crc kubenswrapper[4934]: I1227 07:55:17.289248 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:17 crc kubenswrapper[4934]: I1227 07:55:17.290983 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"01f91788-9f82-4508-906b-1e98c9e05c2c","Type":"ContainerStarted","Data":"3beabe409cdf6e252a6d0cae88496117d87ef061b2377eedda5e43126537661f"} Dec 27 07:55:17 crc kubenswrapper[4934]: I1227 07:55:17.291303 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:17 crc kubenswrapper[4934]: I1227 07:55:17.317218 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" podStartSLOduration=2.409114532 podStartE2EDuration="7.317197321s" podCreationTimestamp="2025-12-27 07:55:10 +0000 UTC" firstStartedPulling="2025-12-27 07:55:11.428624314 +0000 UTC m=+772.249064908" lastFinishedPulling="2025-12-27 07:55:16.336707103 +0000 UTC m=+777.157147697" observedRunningTime="2025-12-27 07:55:17.309160559 +0000 UTC m=+778.129601213" watchObservedRunningTime="2025-12-27 07:55:17.317197321 +0000 UTC m=+778.137637925" Dec 27 07:55:17 crc kubenswrapper[4934]: I1227 07:55:17.333240 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-compactor-0" podStartSLOduration=3.467327128 podStartE2EDuration="7.333222794s" podCreationTimestamp="2025-12-27 07:55:10 +0000 UTC" firstStartedPulling="2025-12-27 07:55:12.524858596 +0000 UTC m=+773.345299190" lastFinishedPulling="2025-12-27 07:55:16.390754262 +0000 UTC m=+777.211194856" observedRunningTime="2025-12-27 07:55:17.331906211 +0000 UTC m=+778.152346815" watchObservedRunningTime="2025-12-27 07:55:17.333222794 +0000 UTC m=+778.153663408" Dec 27 07:55:17 crc kubenswrapper[4934]: I1227 07:55:17.390720 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" podStartSLOduration=2.427330681 podStartE2EDuration="7.39068879s" podCreationTimestamp="2025-12-27 07:55:10 +0000 UTC" firstStartedPulling="2025-12-27 07:55:11.37559874 +0000 UTC m=+772.196039344" lastFinishedPulling="2025-12-27 07:55:16.338956859 +0000 UTC m=+777.159397453" observedRunningTime="2025-12-27 07:55:17.381756816 +0000 UTC m=+778.202197420" watchObservedRunningTime="2025-12-27 07:55:17.39068879 +0000 UTC m=+778.211129394" Dec 27 07:55:17 crc kubenswrapper[4934]: I1227 07:55:17.394438 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-index-gateway-0" podStartSLOduration=3.656523708 podStartE2EDuration="7.394420634s" podCreationTimestamp="2025-12-27 07:55:10 +0000 UTC" firstStartedPulling="2025-12-27 07:55:12.59774562 +0000 UTC m=+773.418186254" lastFinishedPulling="2025-12-27 07:55:16.335642576 +0000 UTC m=+777.156083180" observedRunningTime="2025-12-27 07:55:17.36047368 +0000 UTC m=+778.180914274" watchObservedRunningTime="2025-12-27 07:55:17.394420634 +0000 UTC m=+778.214861288" Dec 27 07:55:17 crc kubenswrapper[4934]: I1227 07:55:17.415523 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" podStartSLOduration=2.791427712 podStartE2EDuration="7.415487204s" podCreationTimestamp="2025-12-27 07:55:10 +0000 UTC" firstStartedPulling="2025-12-27 07:55:11.712422065 +0000 UTC m=+772.532862659" lastFinishedPulling="2025-12-27 07:55:16.336481567 +0000 UTC m=+777.156922151" observedRunningTime="2025-12-27 07:55:17.402330723 +0000 UTC m=+778.222771337" watchObservedRunningTime="2025-12-27 07:55:17.415487204 +0000 UTC m=+778.235927848" Dec 27 07:55:17 crc kubenswrapper[4934]: I1227 07:55:17.434507 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-ingester-0" podStartSLOduration=3.35853326 podStartE2EDuration="7.434487322s" podCreationTimestamp="2025-12-27 07:55:10 +0000 UTC" firstStartedPulling="2025-12-27 07:55:12.230607302 +0000 UTC m=+773.051047896" lastFinishedPulling="2025-12-27 07:55:16.306561354 +0000 UTC m=+777.127001958" observedRunningTime="2025-12-27 07:55:17.425328712 +0000 UTC m=+778.245769306" watchObservedRunningTime="2025-12-27 07:55:17.434487322 +0000 UTC m=+778.254927916" Dec 27 07:55:19 crc kubenswrapper[4934]: I1227 07:55:19.313389 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" event={"ID":"47cbee2e-c325-4a8d-a26d-94bced60ee78","Type":"ContainerStarted","Data":"a6b83e40ff1abc308f10f1a7fc549561f2e46ce3dbfb9863a39e65e9f625df5c"} Dec 27 07:55:19 crc kubenswrapper[4934]: I1227 07:55:19.314240 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:19 crc kubenswrapper[4934]: I1227 07:55:19.316690 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" event={"ID":"eac5b26c-d598-4c68-95b5-e2583b456af9","Type":"ContainerStarted","Data":"a5547294e32c8e3a3f023cc933e519ee890298b3aacb5a6593b619e3f3cee9f3"} Dec 27 07:55:19 crc kubenswrapper[4934]: I1227 07:55:19.317215 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:19 crc kubenswrapper[4934]: I1227 07:55:19.317283 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:19 crc kubenswrapper[4934]: I1227 07:55:19.326703 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:19 crc kubenswrapper[4934]: I1227 07:55:19.334388 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:19 crc kubenswrapper[4934]: I1227 07:55:19.336609 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" Dec 27 07:55:19 crc kubenswrapper[4934]: I1227 07:55:19.353529 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" podStartSLOduration=2.737891596 podStartE2EDuration="9.353508166s" podCreationTimestamp="2025-12-27 07:55:10 +0000 UTC" firstStartedPulling="2025-12-27 07:55:12.012229338 +0000 UTC m=+772.832669922" lastFinishedPulling="2025-12-27 07:55:18.627845898 +0000 UTC m=+779.448286492" observedRunningTime="2025-12-27 07:55:19.351620608 +0000 UTC m=+780.172061252" watchObservedRunningTime="2025-12-27 07:55:19.353508166 +0000 UTC m=+780.173948770" Dec 27 07:55:20 crc kubenswrapper[4934]: I1227 07:55:20.324836 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:20 crc kubenswrapper[4934]: I1227 07:55:20.336950 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" Dec 27 07:55:20 crc kubenswrapper[4934]: I1227 07:55:20.363258 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" podStartSLOduration=4.088578297 podStartE2EDuration="10.363221478s" podCreationTimestamp="2025-12-27 07:55:10 +0000 UTC" firstStartedPulling="2025-12-27 07:55:12.359497375 +0000 UTC m=+773.179937969" lastFinishedPulling="2025-12-27 07:55:18.634140556 +0000 UTC m=+779.454581150" observedRunningTime="2025-12-27 07:55:19.417793433 +0000 UTC m=+780.238234027" watchObservedRunningTime="2025-12-27 07:55:20.363221478 +0000 UTC m=+781.183662102" Dec 27 07:55:31 crc kubenswrapper[4934]: I1227 07:55:31.075187 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" Dec 27 07:55:31 crc kubenswrapper[4934]: I1227 07:55:31.274904 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 07:55:31 crc kubenswrapper[4934]: I1227 07:55:31.992377 4934 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Dec 27 07:55:31 crc kubenswrapper[4934]: I1227 07:55:31.992755 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="642f3213-8063-4ade-800e-4e05ba16d642" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 27 07:55:32 crc kubenswrapper[4934]: I1227 07:55:32.072026 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Dec 27 07:55:32 crc kubenswrapper[4934]: I1227 07:55:32.139738 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 07:55:40 crc kubenswrapper[4934]: I1227 07:55:40.816656 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" Dec 27 07:55:41 crc kubenswrapper[4934]: I1227 07:55:41.987911 4934 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Dec 27 07:55:41 crc kubenswrapper[4934]: I1227 07:55:41.988001 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="642f3213-8063-4ade-800e-4e05ba16d642" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 27 07:55:51 crc kubenswrapper[4934]: I1227 07:55:51.988224 4934 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Dec 27 07:55:51 crc kubenswrapper[4934]: I1227 07:55:51.988749 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="642f3213-8063-4ade-800e-4e05ba16d642" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 27 07:56:01 crc kubenswrapper[4934]: I1227 07:56:01.986283 4934 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Dec 27 07:56:01 crc kubenswrapper[4934]: I1227 07:56:01.987174 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="642f3213-8063-4ade-800e-4e05ba16d642" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 27 07:56:11 crc kubenswrapper[4934]: I1227 07:56:11.984896 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Dec 27 07:56:16 crc kubenswrapper[4934]: I1227 07:56:16.125483 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7m94m"] Dec 27 07:56:16 crc kubenswrapper[4934]: I1227 07:56:16.127770 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7m94m" Dec 27 07:56:16 crc kubenswrapper[4934]: I1227 07:56:16.142540 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7m94m"] Dec 27 07:56:16 crc kubenswrapper[4934]: I1227 07:56:16.201626 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be9d5c37-e6f2-4925-a834-b72e71954b1d-catalog-content\") pod \"certified-operators-7m94m\" (UID: \"be9d5c37-e6f2-4925-a834-b72e71954b1d\") " pod="openshift-marketplace/certified-operators-7m94m" Dec 27 07:56:16 crc kubenswrapper[4934]: I1227 07:56:16.201735 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t8rq\" (UniqueName: \"kubernetes.io/projected/be9d5c37-e6f2-4925-a834-b72e71954b1d-kube-api-access-5t8rq\") pod \"certified-operators-7m94m\" (UID: \"be9d5c37-e6f2-4925-a834-b72e71954b1d\") " pod="openshift-marketplace/certified-operators-7m94m" Dec 27 07:56:16 crc kubenswrapper[4934]: I1227 07:56:16.201854 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be9d5c37-e6f2-4925-a834-b72e71954b1d-utilities\") pod \"certified-operators-7m94m\" (UID: \"be9d5c37-e6f2-4925-a834-b72e71954b1d\") " pod="openshift-marketplace/certified-operators-7m94m" Dec 27 07:56:16 crc kubenswrapper[4934]: I1227 07:56:16.304226 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be9d5c37-e6f2-4925-a834-b72e71954b1d-catalog-content\") pod \"certified-operators-7m94m\" (UID: \"be9d5c37-e6f2-4925-a834-b72e71954b1d\") " pod="openshift-marketplace/certified-operators-7m94m" Dec 27 07:56:16 crc kubenswrapper[4934]: I1227 07:56:16.304402 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t8rq\" (UniqueName: \"kubernetes.io/projected/be9d5c37-e6f2-4925-a834-b72e71954b1d-kube-api-access-5t8rq\") pod \"certified-operators-7m94m\" (UID: \"be9d5c37-e6f2-4925-a834-b72e71954b1d\") " pod="openshift-marketplace/certified-operators-7m94m" Dec 27 07:56:16 crc kubenswrapper[4934]: I1227 07:56:16.304682 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be9d5c37-e6f2-4925-a834-b72e71954b1d-utilities\") pod \"certified-operators-7m94m\" (UID: \"be9d5c37-e6f2-4925-a834-b72e71954b1d\") " pod="openshift-marketplace/certified-operators-7m94m" Dec 27 07:56:16 crc kubenswrapper[4934]: I1227 07:56:16.304745 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be9d5c37-e6f2-4925-a834-b72e71954b1d-catalog-content\") pod \"certified-operators-7m94m\" (UID: \"be9d5c37-e6f2-4925-a834-b72e71954b1d\") " pod="openshift-marketplace/certified-operators-7m94m" Dec 27 07:56:16 crc kubenswrapper[4934]: I1227 07:56:16.305416 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be9d5c37-e6f2-4925-a834-b72e71954b1d-utilities\") pod \"certified-operators-7m94m\" (UID: \"be9d5c37-e6f2-4925-a834-b72e71954b1d\") " pod="openshift-marketplace/certified-operators-7m94m" Dec 27 07:56:16 crc kubenswrapper[4934]: I1227 07:56:16.325480 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t8rq\" (UniqueName: \"kubernetes.io/projected/be9d5c37-e6f2-4925-a834-b72e71954b1d-kube-api-access-5t8rq\") pod \"certified-operators-7m94m\" (UID: \"be9d5c37-e6f2-4925-a834-b72e71954b1d\") " pod="openshift-marketplace/certified-operators-7m94m" Dec 27 07:56:16 crc kubenswrapper[4934]: I1227 07:56:16.445015 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7m94m" Dec 27 07:56:16 crc kubenswrapper[4934]: I1227 07:56:16.932987 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7m94m"] Dec 27 07:56:17 crc kubenswrapper[4934]: I1227 07:56:17.856723 4934 generic.go:334] "Generic (PLEG): container finished" podID="be9d5c37-e6f2-4925-a834-b72e71954b1d" containerID="e18772b33c9dd2926961db5d48961909b683e60d4b11b8cc70150a71ff730ab3" exitCode=0 Dec 27 07:56:17 crc kubenswrapper[4934]: I1227 07:56:17.856809 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7m94m" event={"ID":"be9d5c37-e6f2-4925-a834-b72e71954b1d","Type":"ContainerDied","Data":"e18772b33c9dd2926961db5d48961909b683e60d4b11b8cc70150a71ff730ab3"} Dec 27 07:56:17 crc kubenswrapper[4934]: I1227 07:56:17.857123 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7m94m" event={"ID":"be9d5c37-e6f2-4925-a834-b72e71954b1d","Type":"ContainerStarted","Data":"aef1e8beda66bec9a1ddb6c0bb841a32dfc51fb57404b3c34444b789277c0fce"} Dec 27 07:56:18 crc kubenswrapper[4934]: I1227 07:56:18.869811 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7m94m" event={"ID":"be9d5c37-e6f2-4925-a834-b72e71954b1d","Type":"ContainerStarted","Data":"d49821120f4a5ebf75efddf62291b65a9a7fb3ba3c7cc40e9b03bb0ec2e68d6a"} Dec 27 07:56:19 crc kubenswrapper[4934]: I1227 07:56:19.881696 4934 generic.go:334] "Generic (PLEG): container finished" podID="be9d5c37-e6f2-4925-a834-b72e71954b1d" containerID="d49821120f4a5ebf75efddf62291b65a9a7fb3ba3c7cc40e9b03bb0ec2e68d6a" exitCode=0 Dec 27 07:56:19 crc kubenswrapper[4934]: I1227 07:56:19.881857 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7m94m" event={"ID":"be9d5c37-e6f2-4925-a834-b72e71954b1d","Type":"ContainerDied","Data":"d49821120f4a5ebf75efddf62291b65a9a7fb3ba3c7cc40e9b03bb0ec2e68d6a"} Dec 27 07:56:20 crc kubenswrapper[4934]: I1227 07:56:20.896501 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7m94m" event={"ID":"be9d5c37-e6f2-4925-a834-b72e71954b1d","Type":"ContainerStarted","Data":"ebb3f7e50398081371bd87bdaa2cd34244b414846fc0086bf7c6d4732574e588"} Dec 27 07:56:20 crc kubenswrapper[4934]: I1227 07:56:20.931775 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7m94m" podStartSLOduration=2.484001647 podStartE2EDuration="4.9317528s" podCreationTimestamp="2025-12-27 07:56:16 +0000 UTC" firstStartedPulling="2025-12-27 07:56:17.858725481 +0000 UTC m=+838.679166075" lastFinishedPulling="2025-12-27 07:56:20.306476594 +0000 UTC m=+841.126917228" observedRunningTime="2025-12-27 07:56:20.923805676 +0000 UTC m=+841.744246300" watchObservedRunningTime="2025-12-27 07:56:20.9317528 +0000 UTC m=+841.752193414" Dec 27 07:56:26 crc kubenswrapper[4934]: I1227 07:56:26.445300 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7m94m" Dec 27 07:56:26 crc kubenswrapper[4934]: I1227 07:56:26.445741 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7m94m" Dec 27 07:56:26 crc kubenswrapper[4934]: I1227 07:56:26.516411 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7m94m" Dec 27 07:56:26 crc kubenswrapper[4934]: I1227 07:56:26.978765 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7m94m" Dec 27 07:56:27 crc kubenswrapper[4934]: I1227 07:56:27.033022 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7m94m"] Dec 27 07:56:28 crc kubenswrapper[4934]: I1227 07:56:28.974519 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7m94m" podUID="be9d5c37-e6f2-4925-a834-b72e71954b1d" containerName="registry-server" containerID="cri-o://ebb3f7e50398081371bd87bdaa2cd34244b414846fc0086bf7c6d4732574e588" gracePeriod=2 Dec 27 07:56:29 crc kubenswrapper[4934]: I1227 07:56:29.626494 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7m94m" Dec 27 07:56:29 crc kubenswrapper[4934]: I1227 07:56:29.740954 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be9d5c37-e6f2-4925-a834-b72e71954b1d-catalog-content\") pod \"be9d5c37-e6f2-4925-a834-b72e71954b1d\" (UID: \"be9d5c37-e6f2-4925-a834-b72e71954b1d\") " Dec 27 07:56:29 crc kubenswrapper[4934]: I1227 07:56:29.741124 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5t8rq\" (UniqueName: \"kubernetes.io/projected/be9d5c37-e6f2-4925-a834-b72e71954b1d-kube-api-access-5t8rq\") pod \"be9d5c37-e6f2-4925-a834-b72e71954b1d\" (UID: \"be9d5c37-e6f2-4925-a834-b72e71954b1d\") " Dec 27 07:56:29 crc kubenswrapper[4934]: I1227 07:56:29.741187 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be9d5c37-e6f2-4925-a834-b72e71954b1d-utilities\") pod \"be9d5c37-e6f2-4925-a834-b72e71954b1d\" (UID: \"be9d5c37-e6f2-4925-a834-b72e71954b1d\") " Dec 27 07:56:29 crc kubenswrapper[4934]: I1227 07:56:29.742070 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be9d5c37-e6f2-4925-a834-b72e71954b1d-utilities" (OuterVolumeSpecName: "utilities") pod "be9d5c37-e6f2-4925-a834-b72e71954b1d" (UID: "be9d5c37-e6f2-4925-a834-b72e71954b1d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:56:29 crc kubenswrapper[4934]: I1227 07:56:29.759396 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be9d5c37-e6f2-4925-a834-b72e71954b1d-kube-api-access-5t8rq" (OuterVolumeSpecName: "kube-api-access-5t8rq") pod "be9d5c37-e6f2-4925-a834-b72e71954b1d" (UID: "be9d5c37-e6f2-4925-a834-b72e71954b1d"). InnerVolumeSpecName "kube-api-access-5t8rq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:56:29 crc kubenswrapper[4934]: I1227 07:56:29.791919 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be9d5c37-e6f2-4925-a834-b72e71954b1d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be9d5c37-e6f2-4925-a834-b72e71954b1d" (UID: "be9d5c37-e6f2-4925-a834-b72e71954b1d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:56:29 crc kubenswrapper[4934]: I1227 07:56:29.843323 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be9d5c37-e6f2-4925-a834-b72e71954b1d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 07:56:29 crc kubenswrapper[4934]: I1227 07:56:29.843356 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5t8rq\" (UniqueName: \"kubernetes.io/projected/be9d5c37-e6f2-4925-a834-b72e71954b1d-kube-api-access-5t8rq\") on node \"crc\" DevicePath \"\"" Dec 27 07:56:29 crc kubenswrapper[4934]: I1227 07:56:29.843366 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be9d5c37-e6f2-4925-a834-b72e71954b1d-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 07:56:29 crc kubenswrapper[4934]: I1227 07:56:29.982622 4934 generic.go:334] "Generic (PLEG): container finished" podID="be9d5c37-e6f2-4925-a834-b72e71954b1d" containerID="ebb3f7e50398081371bd87bdaa2cd34244b414846fc0086bf7c6d4732574e588" exitCode=0 Dec 27 07:56:29 crc kubenswrapper[4934]: I1227 07:56:29.982679 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7m94m" event={"ID":"be9d5c37-e6f2-4925-a834-b72e71954b1d","Type":"ContainerDied","Data":"ebb3f7e50398081371bd87bdaa2cd34244b414846fc0086bf7c6d4732574e588"} Dec 27 07:56:29 crc kubenswrapper[4934]: I1227 07:56:29.982718 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7m94m" event={"ID":"be9d5c37-e6f2-4925-a834-b72e71954b1d","Type":"ContainerDied","Data":"aef1e8beda66bec9a1ddb6c0bb841a32dfc51fb57404b3c34444b789277c0fce"} Dec 27 07:56:29 crc kubenswrapper[4934]: I1227 07:56:29.982743 4934 scope.go:117] "RemoveContainer" containerID="ebb3f7e50398081371bd87bdaa2cd34244b414846fc0086bf7c6d4732574e588" Dec 27 07:56:29 crc kubenswrapper[4934]: I1227 07:56:29.982683 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7m94m" Dec 27 07:56:30 crc kubenswrapper[4934]: I1227 07:56:30.010544 4934 scope.go:117] "RemoveContainer" containerID="d49821120f4a5ebf75efddf62291b65a9a7fb3ba3c7cc40e9b03bb0ec2e68d6a" Dec 27 07:56:30 crc kubenswrapper[4934]: I1227 07:56:30.022119 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7m94m"] Dec 27 07:56:30 crc kubenswrapper[4934]: I1227 07:56:30.030740 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7m94m"] Dec 27 07:56:30 crc kubenswrapper[4934]: I1227 07:56:30.057618 4934 scope.go:117] "RemoveContainer" containerID="e18772b33c9dd2926961db5d48961909b683e60d4b11b8cc70150a71ff730ab3" Dec 27 07:56:30 crc kubenswrapper[4934]: I1227 07:56:30.080265 4934 scope.go:117] "RemoveContainer" containerID="ebb3f7e50398081371bd87bdaa2cd34244b414846fc0086bf7c6d4732574e588" Dec 27 07:56:30 crc kubenswrapper[4934]: E1227 07:56:30.080675 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebb3f7e50398081371bd87bdaa2cd34244b414846fc0086bf7c6d4732574e588\": container with ID starting with ebb3f7e50398081371bd87bdaa2cd34244b414846fc0086bf7c6d4732574e588 not found: ID does not exist" containerID="ebb3f7e50398081371bd87bdaa2cd34244b414846fc0086bf7c6d4732574e588" Dec 27 07:56:30 crc kubenswrapper[4934]: I1227 07:56:30.080728 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebb3f7e50398081371bd87bdaa2cd34244b414846fc0086bf7c6d4732574e588"} err="failed to get container status \"ebb3f7e50398081371bd87bdaa2cd34244b414846fc0086bf7c6d4732574e588\": rpc error: code = NotFound desc = could not find container \"ebb3f7e50398081371bd87bdaa2cd34244b414846fc0086bf7c6d4732574e588\": container with ID starting with ebb3f7e50398081371bd87bdaa2cd34244b414846fc0086bf7c6d4732574e588 not found: ID does not exist" Dec 27 07:56:30 crc kubenswrapper[4934]: I1227 07:56:30.080767 4934 scope.go:117] "RemoveContainer" containerID="d49821120f4a5ebf75efddf62291b65a9a7fb3ba3c7cc40e9b03bb0ec2e68d6a" Dec 27 07:56:30 crc kubenswrapper[4934]: E1227 07:56:30.081140 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d49821120f4a5ebf75efddf62291b65a9a7fb3ba3c7cc40e9b03bb0ec2e68d6a\": container with ID starting with d49821120f4a5ebf75efddf62291b65a9a7fb3ba3c7cc40e9b03bb0ec2e68d6a not found: ID does not exist" containerID="d49821120f4a5ebf75efddf62291b65a9a7fb3ba3c7cc40e9b03bb0ec2e68d6a" Dec 27 07:56:30 crc kubenswrapper[4934]: I1227 07:56:30.081171 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d49821120f4a5ebf75efddf62291b65a9a7fb3ba3c7cc40e9b03bb0ec2e68d6a"} err="failed to get container status \"d49821120f4a5ebf75efddf62291b65a9a7fb3ba3c7cc40e9b03bb0ec2e68d6a\": rpc error: code = NotFound desc = could not find container \"d49821120f4a5ebf75efddf62291b65a9a7fb3ba3c7cc40e9b03bb0ec2e68d6a\": container with ID starting with d49821120f4a5ebf75efddf62291b65a9a7fb3ba3c7cc40e9b03bb0ec2e68d6a not found: ID does not exist" Dec 27 07:56:30 crc kubenswrapper[4934]: I1227 07:56:30.081186 4934 scope.go:117] "RemoveContainer" containerID="e18772b33c9dd2926961db5d48961909b683e60d4b11b8cc70150a71ff730ab3" Dec 27 07:56:30 crc kubenswrapper[4934]: E1227 07:56:30.081537 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e18772b33c9dd2926961db5d48961909b683e60d4b11b8cc70150a71ff730ab3\": container with ID starting with e18772b33c9dd2926961db5d48961909b683e60d4b11b8cc70150a71ff730ab3 not found: ID does not exist" containerID="e18772b33c9dd2926961db5d48961909b683e60d4b11b8cc70150a71ff730ab3" Dec 27 07:56:30 crc kubenswrapper[4934]: I1227 07:56:30.081557 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e18772b33c9dd2926961db5d48961909b683e60d4b11b8cc70150a71ff730ab3"} err="failed to get container status \"e18772b33c9dd2926961db5d48961909b683e60d4b11b8cc70150a71ff730ab3\": rpc error: code = NotFound desc = could not find container \"e18772b33c9dd2926961db5d48961909b683e60d4b11b8cc70150a71ff730ab3\": container with ID starting with e18772b33c9dd2926961db5d48961909b683e60d4b11b8cc70150a71ff730ab3 not found: ID does not exist" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.186479 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-r6vwb"] Dec 27 07:56:31 crc kubenswrapper[4934]: E1227 07:56:31.186936 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be9d5c37-e6f2-4925-a834-b72e71954b1d" containerName="extract-utilities" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.186953 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="be9d5c37-e6f2-4925-a834-b72e71954b1d" containerName="extract-utilities" Dec 27 07:56:31 crc kubenswrapper[4934]: E1227 07:56:31.186983 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be9d5c37-e6f2-4925-a834-b72e71954b1d" containerName="extract-content" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.186992 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="be9d5c37-e6f2-4925-a834-b72e71954b1d" containerName="extract-content" Dec 27 07:56:31 crc kubenswrapper[4934]: E1227 07:56:31.187002 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be9d5c37-e6f2-4925-a834-b72e71954b1d" containerName="registry-server" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.187008 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="be9d5c37-e6f2-4925-a834-b72e71954b1d" containerName="registry-server" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.187161 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="be9d5c37-e6f2-4925-a834-b72e71954b1d" containerName="registry-server" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.187853 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.191324 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-z7shv" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.191777 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.191801 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.191880 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.193928 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.213472 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.216101 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-r6vwb"] Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.269225 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-collector-token\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.269563 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-trusted-ca\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.269750 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-config-openshift-service-cacrt\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.269905 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-config\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.270106 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/f9cb51ea-e813-433f-b678-e0e4cee8b856-datadir\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.270267 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-entrypoint\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.270457 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-collector-syslog-receiver\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.270640 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/f9cb51ea-e813-433f-b678-e0e4cee8b856-tmp\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.270789 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/f9cb51ea-e813-433f-b678-e0e4cee8b856-sa-token\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.270930 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-metrics\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.271108 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4t2m\" (UniqueName: \"kubernetes.io/projected/f9cb51ea-e813-433f-b678-e0e4cee8b856-kube-api-access-j4t2m\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.345865 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-r6vwb"] Dec 27 07:56:31 crc kubenswrapper[4934]: E1227 07:56:31.347451 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[collector-syslog-receiver collector-token config config-openshift-service-cacrt datadir entrypoint kube-api-access-j4t2m metrics sa-token tmp trusted-ca], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-logging/collector-r6vwb" podUID="f9cb51ea-e813-433f-b678-e0e4cee8b856" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.372935 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/f9cb51ea-e813-433f-b678-e0e4cee8b856-tmp\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.372980 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/f9cb51ea-e813-433f-b678-e0e4cee8b856-sa-token\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.373000 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-metrics\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.373024 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4t2m\" (UniqueName: \"kubernetes.io/projected/f9cb51ea-e813-433f-b678-e0e4cee8b856-kube-api-access-j4t2m\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.373055 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-collector-token\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.373102 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-trusted-ca\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.373137 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-config-openshift-service-cacrt\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.373157 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-config\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.373186 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/f9cb51ea-e813-433f-b678-e0e4cee8b856-datadir\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.373211 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-entrypoint\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.373231 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-collector-syslog-receiver\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.373330 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/f9cb51ea-e813-433f-b678-e0e4cee8b856-datadir\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: E1227 07:56:31.373359 4934 secret.go:188] Couldn't get secret openshift-logging/collector-syslog-receiver: secret "collector-syslog-receiver" not found Dec 27 07:56:31 crc kubenswrapper[4934]: E1227 07:56:31.373434 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-collector-syslog-receiver podName:f9cb51ea-e813-433f-b678-e0e4cee8b856 nodeName:}" failed. No retries permitted until 2025-12-27 07:56:31.873416162 +0000 UTC m=+852.693856756 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "collector-syslog-receiver" (UniqueName: "kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-collector-syslog-receiver") pod "collector-r6vwb" (UID: "f9cb51ea-e813-433f-b678-e0e4cee8b856") : secret "collector-syslog-receiver" not found Dec 27 07:56:31 crc kubenswrapper[4934]: E1227 07:56:31.373547 4934 secret.go:188] Couldn't get secret openshift-logging/collector-metrics: secret "collector-metrics" not found Dec 27 07:56:31 crc kubenswrapper[4934]: E1227 07:56:31.373606 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-metrics podName:f9cb51ea-e813-433f-b678-e0e4cee8b856 nodeName:}" failed. No retries permitted until 2025-12-27 07:56:31.873588657 +0000 UTC m=+852.694029251 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics" (UniqueName: "kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-metrics") pod "collector-r6vwb" (UID: "f9cb51ea-e813-433f-b678-e0e4cee8b856") : secret "collector-metrics" not found Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.373970 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-config\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.374021 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-config-openshift-service-cacrt\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.374137 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-entrypoint\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.374170 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-trusted-ca\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.377717 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-collector-token\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.389386 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/f9cb51ea-e813-433f-b678-e0e4cee8b856-tmp\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.390381 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/f9cb51ea-e813-433f-b678-e0e4cee8b856-sa-token\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.395495 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4t2m\" (UniqueName: \"kubernetes.io/projected/f9cb51ea-e813-433f-b678-e0e4cee8b856-kube-api-access-j4t2m\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.478784 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be9d5c37-e6f2-4925-a834-b72e71954b1d" path="/var/lib/kubelet/pods/be9d5c37-e6f2-4925-a834-b72e71954b1d/volumes" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.884440 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-collector-syslog-receiver\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.884584 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-metrics\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.889257 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-collector-syslog-receiver\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:31 crc kubenswrapper[4934]: I1227 07:56:31.889388 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-metrics\") pod \"collector-r6vwb\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " pod="openshift-logging/collector-r6vwb" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.011175 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-r6vwb" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.022987 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-r6vwb" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.088517 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/f9cb51ea-e813-433f-b678-e0e4cee8b856-datadir\") pod \"f9cb51ea-e813-433f-b678-e0e4cee8b856\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.088584 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/f9cb51ea-e813-433f-b678-e0e4cee8b856-sa-token\") pod \"f9cb51ea-e813-433f-b678-e0e4cee8b856\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.088610 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f9cb51ea-e813-433f-b678-e0e4cee8b856-datadir" (OuterVolumeSpecName: "datadir") pod "f9cb51ea-e813-433f-b678-e0e4cee8b856" (UID: "f9cb51ea-e813-433f-b678-e0e4cee8b856"). InnerVolumeSpecName "datadir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.088636 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-collector-token\") pod \"f9cb51ea-e813-433f-b678-e0e4cee8b856\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.088741 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-metrics\") pod \"f9cb51ea-e813-433f-b678-e0e4cee8b856\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.088796 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-config-openshift-service-cacrt\") pod \"f9cb51ea-e813-433f-b678-e0e4cee8b856\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.088839 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-collector-syslog-receiver\") pod \"f9cb51ea-e813-433f-b678-e0e4cee8b856\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.088883 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-config\") pod \"f9cb51ea-e813-433f-b678-e0e4cee8b856\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.088912 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-trusted-ca\") pod \"f9cb51ea-e813-433f-b678-e0e4cee8b856\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.088955 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/f9cb51ea-e813-433f-b678-e0e4cee8b856-tmp\") pod \"f9cb51ea-e813-433f-b678-e0e4cee8b856\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.088991 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4t2m\" (UniqueName: \"kubernetes.io/projected/f9cb51ea-e813-433f-b678-e0e4cee8b856-kube-api-access-j4t2m\") pod \"f9cb51ea-e813-433f-b678-e0e4cee8b856\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.089056 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-entrypoint\") pod \"f9cb51ea-e813-433f-b678-e0e4cee8b856\" (UID: \"f9cb51ea-e813-433f-b678-e0e4cee8b856\") " Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.089532 4934 reconciler_common.go:293] "Volume detached for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/f9cb51ea-e813-433f-b678-e0e4cee8b856-datadir\") on node \"crc\" DevicePath \"\"" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.090127 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-config-openshift-service-cacrt" (OuterVolumeSpecName: "config-openshift-service-cacrt") pod "f9cb51ea-e813-433f-b678-e0e4cee8b856" (UID: "f9cb51ea-e813-433f-b678-e0e4cee8b856"). InnerVolumeSpecName "config-openshift-service-cacrt". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.090294 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-entrypoint" (OuterVolumeSpecName: "entrypoint") pod "f9cb51ea-e813-433f-b678-e0e4cee8b856" (UID: "f9cb51ea-e813-433f-b678-e0e4cee8b856"). InnerVolumeSpecName "entrypoint". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.090316 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-config" (OuterVolumeSpecName: "config") pod "f9cb51ea-e813-433f-b678-e0e4cee8b856" (UID: "f9cb51ea-e813-433f-b678-e0e4cee8b856"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.090708 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "f9cb51ea-e813-433f-b678-e0e4cee8b856" (UID: "f9cb51ea-e813-433f-b678-e0e4cee8b856"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.093068 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-collector-syslog-receiver" (OuterVolumeSpecName: "collector-syslog-receiver") pod "f9cb51ea-e813-433f-b678-e0e4cee8b856" (UID: "f9cb51ea-e813-433f-b678-e0e4cee8b856"). InnerVolumeSpecName "collector-syslog-receiver". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.093324 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-metrics" (OuterVolumeSpecName: "metrics") pod "f9cb51ea-e813-433f-b678-e0e4cee8b856" (UID: "f9cb51ea-e813-433f-b678-e0e4cee8b856"). InnerVolumeSpecName "metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.093850 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9cb51ea-e813-433f-b678-e0e4cee8b856-kube-api-access-j4t2m" (OuterVolumeSpecName: "kube-api-access-j4t2m") pod "f9cb51ea-e813-433f-b678-e0e4cee8b856" (UID: "f9cb51ea-e813-433f-b678-e0e4cee8b856"). InnerVolumeSpecName "kube-api-access-j4t2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.094019 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9cb51ea-e813-433f-b678-e0e4cee8b856-sa-token" (OuterVolumeSpecName: "sa-token") pod "f9cb51ea-e813-433f-b678-e0e4cee8b856" (UID: "f9cb51ea-e813-433f-b678-e0e4cee8b856"). InnerVolumeSpecName "sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.094405 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9cb51ea-e813-433f-b678-e0e4cee8b856-tmp" (OuterVolumeSpecName: "tmp") pod "f9cb51ea-e813-433f-b678-e0e4cee8b856" (UID: "f9cb51ea-e813-433f-b678-e0e4cee8b856"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.094723 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-collector-token" (OuterVolumeSpecName: "collector-token") pod "f9cb51ea-e813-433f-b678-e0e4cee8b856" (UID: "f9cb51ea-e813-433f-b678-e0e4cee8b856"). InnerVolumeSpecName "collector-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.190444 4934 reconciler_common.go:293] "Volume detached for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-metrics\") on node \"crc\" DevicePath \"\"" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.190526 4934 reconciler_common.go:293] "Volume detached for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-config-openshift-service-cacrt\") on node \"crc\" DevicePath \"\"" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.190548 4934 reconciler_common.go:293] "Volume detached for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-collector-syslog-receiver\") on node \"crc\" DevicePath \"\"" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.190567 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.190585 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.190602 4934 reconciler_common.go:293] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/f9cb51ea-e813-433f-b678-e0e4cee8b856-tmp\") on node \"crc\" DevicePath \"\"" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.190618 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4t2m\" (UniqueName: \"kubernetes.io/projected/f9cb51ea-e813-433f-b678-e0e4cee8b856-kube-api-access-j4t2m\") on node \"crc\" DevicePath \"\"" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.190634 4934 reconciler_common.go:293] "Volume detached for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/f9cb51ea-e813-433f-b678-e0e4cee8b856-entrypoint\") on node \"crc\" DevicePath \"\"" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.190651 4934 reconciler_common.go:293] "Volume detached for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/f9cb51ea-e813-433f-b678-e0e4cee8b856-sa-token\") on node \"crc\" DevicePath \"\"" Dec 27 07:56:32 crc kubenswrapper[4934]: I1227 07:56:32.190669 4934 reconciler_common.go:293] "Volume detached for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/f9cb51ea-e813-433f-b678-e0e4cee8b856-collector-token\") on node \"crc\" DevicePath \"\"" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.021226 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-r6vwb" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.110216 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-r6vwb"] Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.131961 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-logging/collector-r6vwb"] Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.139716 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-dngrf"] Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.141835 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.146883 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-dngrf"] Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.149406 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.150152 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.150888 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.151020 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.151191 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-z7shv" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.156816 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.213224 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/dc561c93-57e5-4231-9f5d-6b93101c4b38-entrypoint\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.213288 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/dc561c93-57e5-4231-9f5d-6b93101c4b38-config-openshift-service-cacrt\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.213355 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/dc561c93-57e5-4231-9f5d-6b93101c4b38-sa-token\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.213390 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/dc561c93-57e5-4231-9f5d-6b93101c4b38-tmp\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.213432 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc561c93-57e5-4231-9f5d-6b93101c4b38-config\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.213456 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc561c93-57e5-4231-9f5d-6b93101c4b38-trusted-ca\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.213482 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/dc561c93-57e5-4231-9f5d-6b93101c4b38-metrics\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.213518 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/dc561c93-57e5-4231-9f5d-6b93101c4b38-collector-token\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.213563 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/dc561c93-57e5-4231-9f5d-6b93101c4b38-collector-syslog-receiver\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.213612 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/dc561c93-57e5-4231-9f5d-6b93101c4b38-datadir\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.213661 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5rvg\" (UniqueName: \"kubernetes.io/projected/dc561c93-57e5-4231-9f5d-6b93101c4b38-kube-api-access-s5rvg\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.315066 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc561c93-57e5-4231-9f5d-6b93101c4b38-config\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.315166 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc561c93-57e5-4231-9f5d-6b93101c4b38-trusted-ca\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.315200 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/dc561c93-57e5-4231-9f5d-6b93101c4b38-metrics\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.315229 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/dc561c93-57e5-4231-9f5d-6b93101c4b38-collector-token\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.315279 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/dc561c93-57e5-4231-9f5d-6b93101c4b38-collector-syslog-receiver\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.315328 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/dc561c93-57e5-4231-9f5d-6b93101c4b38-datadir\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.315352 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5rvg\" (UniqueName: \"kubernetes.io/projected/dc561c93-57e5-4231-9f5d-6b93101c4b38-kube-api-access-s5rvg\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.315390 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/dc561c93-57e5-4231-9f5d-6b93101c4b38-entrypoint\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.315430 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/dc561c93-57e5-4231-9f5d-6b93101c4b38-config-openshift-service-cacrt\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.315463 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/dc561c93-57e5-4231-9f5d-6b93101c4b38-sa-token\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.315491 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/dc561c93-57e5-4231-9f5d-6b93101c4b38-tmp\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.316807 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/dc561c93-57e5-4231-9f5d-6b93101c4b38-config-openshift-service-cacrt\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.316967 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/dc561c93-57e5-4231-9f5d-6b93101c4b38-entrypoint\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.317228 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dc561c93-57e5-4231-9f5d-6b93101c4b38-trusted-ca\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.317647 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/dc561c93-57e5-4231-9f5d-6b93101c4b38-datadir\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.318014 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc561c93-57e5-4231-9f5d-6b93101c4b38-config\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.320119 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/dc561c93-57e5-4231-9f5d-6b93101c4b38-tmp\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.321801 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/dc561c93-57e5-4231-9f5d-6b93101c4b38-metrics\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.322118 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/dc561c93-57e5-4231-9f5d-6b93101c4b38-collector-token\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.325499 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/dc561c93-57e5-4231-9f5d-6b93101c4b38-collector-syslog-receiver\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.347245 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/dc561c93-57e5-4231-9f5d-6b93101c4b38-sa-token\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.348046 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5rvg\" (UniqueName: \"kubernetes.io/projected/dc561c93-57e5-4231-9f5d-6b93101c4b38-kube-api-access-s5rvg\") pod \"collector-dngrf\" (UID: \"dc561c93-57e5-4231-9f5d-6b93101c4b38\") " pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.473421 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-dngrf" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.483381 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9cb51ea-e813-433f-b678-e0e4cee8b856" path="/var/lib/kubelet/pods/f9cb51ea-e813-433f-b678-e0e4cee8b856/volumes" Dec 27 07:56:33 crc kubenswrapper[4934]: I1227 07:56:33.995513 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-dngrf"] Dec 27 07:56:34 crc kubenswrapper[4934]: I1227 07:56:34.032886 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-dngrf" event={"ID":"dc561c93-57e5-4231-9f5d-6b93101c4b38","Type":"ContainerStarted","Data":"b8421f552e9464dd69aac8a147b9c6572680a7407eee79735b0510070cfff10b"} Dec 27 07:56:40 crc kubenswrapper[4934]: I1227 07:56:40.087299 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-dngrf" event={"ID":"dc561c93-57e5-4231-9f5d-6b93101c4b38","Type":"ContainerStarted","Data":"aff9137ff5f755737a8beefbeb7afb91dde641166f2ca1b3c25b927af42b75f5"} Dec 27 07:57:15 crc kubenswrapper[4934]: I1227 07:57:15.895490 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/collector-dngrf" podStartSLOduration=37.194369121 podStartE2EDuration="42.895469151s" podCreationTimestamp="2025-12-27 07:56:33 +0000 UTC" firstStartedPulling="2025-12-27 07:56:34.006293361 +0000 UTC m=+854.826733985" lastFinishedPulling="2025-12-27 07:56:39.707393421 +0000 UTC m=+860.527834015" observedRunningTime="2025-12-27 07:56:40.121908058 +0000 UTC m=+860.942348673" watchObservedRunningTime="2025-12-27 07:57:15.895469151 +0000 UTC m=+896.715909755" Dec 27 07:57:15 crc kubenswrapper[4934]: I1227 07:57:15.897445 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp"] Dec 27 07:57:15 crc kubenswrapper[4934]: I1227 07:57:15.899058 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp" Dec 27 07:57:15 crc kubenswrapper[4934]: I1227 07:57:15.901117 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 27 07:57:15 crc kubenswrapper[4934]: I1227 07:57:15.905725 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp"] Dec 27 07:57:16 crc kubenswrapper[4934]: I1227 07:57:16.045762 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sd2s\" (UniqueName: \"kubernetes.io/projected/7f133bd4-74ca-4b6c-bd9a-210102808e22-kube-api-access-8sd2s\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp\" (UID: \"7f133bd4-74ca-4b6c-bd9a-210102808e22\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp" Dec 27 07:57:16 crc kubenswrapper[4934]: I1227 07:57:16.045852 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7f133bd4-74ca-4b6c-bd9a-210102808e22-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp\" (UID: \"7f133bd4-74ca-4b6c-bd9a-210102808e22\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp" Dec 27 07:57:16 crc kubenswrapper[4934]: I1227 07:57:16.045889 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7f133bd4-74ca-4b6c-bd9a-210102808e22-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp\" (UID: \"7f133bd4-74ca-4b6c-bd9a-210102808e22\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp" Dec 27 07:57:16 crc kubenswrapper[4934]: I1227 07:57:16.147876 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sd2s\" (UniqueName: \"kubernetes.io/projected/7f133bd4-74ca-4b6c-bd9a-210102808e22-kube-api-access-8sd2s\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp\" (UID: \"7f133bd4-74ca-4b6c-bd9a-210102808e22\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp" Dec 27 07:57:16 crc kubenswrapper[4934]: I1227 07:57:16.147960 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7f133bd4-74ca-4b6c-bd9a-210102808e22-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp\" (UID: \"7f133bd4-74ca-4b6c-bd9a-210102808e22\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp" Dec 27 07:57:16 crc kubenswrapper[4934]: I1227 07:57:16.147995 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7f133bd4-74ca-4b6c-bd9a-210102808e22-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp\" (UID: \"7f133bd4-74ca-4b6c-bd9a-210102808e22\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp" Dec 27 07:57:16 crc kubenswrapper[4934]: I1227 07:57:16.148626 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7f133bd4-74ca-4b6c-bd9a-210102808e22-util\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp\" (UID: \"7f133bd4-74ca-4b6c-bd9a-210102808e22\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp" Dec 27 07:57:16 crc kubenswrapper[4934]: I1227 07:57:16.148679 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7f133bd4-74ca-4b6c-bd9a-210102808e22-bundle\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp\" (UID: \"7f133bd4-74ca-4b6c-bd9a-210102808e22\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp" Dec 27 07:57:16 crc kubenswrapper[4934]: I1227 07:57:16.168048 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sd2s\" (UniqueName: \"kubernetes.io/projected/7f133bd4-74ca-4b6c-bd9a-210102808e22-kube-api-access-8sd2s\") pod \"98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp\" (UID: \"7f133bd4-74ca-4b6c-bd9a-210102808e22\") " pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp" Dec 27 07:57:16 crc kubenswrapper[4934]: I1227 07:57:16.220601 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp" Dec 27 07:57:16 crc kubenswrapper[4934]: I1227 07:57:16.666878 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp"] Dec 27 07:57:17 crc kubenswrapper[4934]: I1227 07:57:17.412439 4934 generic.go:334] "Generic (PLEG): container finished" podID="7f133bd4-74ca-4b6c-bd9a-210102808e22" containerID="22884de3329d8ba1cca03fc442c2a3bab9f8b1a9a58345800bd0ea573fcf8574" exitCode=0 Dec 27 07:57:17 crc kubenswrapper[4934]: I1227 07:57:17.412484 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp" event={"ID":"7f133bd4-74ca-4b6c-bd9a-210102808e22","Type":"ContainerDied","Data":"22884de3329d8ba1cca03fc442c2a3bab9f8b1a9a58345800bd0ea573fcf8574"} Dec 27 07:57:17 crc kubenswrapper[4934]: I1227 07:57:17.412762 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp" event={"ID":"7f133bd4-74ca-4b6c-bd9a-210102808e22","Type":"ContainerStarted","Data":"b53feb2edd2913969b2adde2698344d52d0010c2d6f5f8224229f66a83953376"} Dec 27 07:57:18 crc kubenswrapper[4934]: I1227 07:57:18.261006 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wc8dz"] Dec 27 07:57:18 crc kubenswrapper[4934]: I1227 07:57:18.263205 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wc8dz" Dec 27 07:57:18 crc kubenswrapper[4934]: I1227 07:57:18.275940 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wc8dz"] Dec 27 07:57:18 crc kubenswrapper[4934]: I1227 07:57:18.388662 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81f8d777-8d08-49e0-b3c2-5ef2cb2388e3-catalog-content\") pod \"redhat-operators-wc8dz\" (UID: \"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3\") " pod="openshift-marketplace/redhat-operators-wc8dz" Dec 27 07:57:18 crc kubenswrapper[4934]: I1227 07:57:18.389040 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw2v2\" (UniqueName: \"kubernetes.io/projected/81f8d777-8d08-49e0-b3c2-5ef2cb2388e3-kube-api-access-vw2v2\") pod \"redhat-operators-wc8dz\" (UID: \"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3\") " pod="openshift-marketplace/redhat-operators-wc8dz" Dec 27 07:57:18 crc kubenswrapper[4934]: I1227 07:57:18.389387 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81f8d777-8d08-49e0-b3c2-5ef2cb2388e3-utilities\") pod \"redhat-operators-wc8dz\" (UID: \"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3\") " pod="openshift-marketplace/redhat-operators-wc8dz" Dec 27 07:57:18 crc kubenswrapper[4934]: I1227 07:57:18.490491 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw2v2\" (UniqueName: \"kubernetes.io/projected/81f8d777-8d08-49e0-b3c2-5ef2cb2388e3-kube-api-access-vw2v2\") pod \"redhat-operators-wc8dz\" (UID: \"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3\") " pod="openshift-marketplace/redhat-operators-wc8dz" Dec 27 07:57:18 crc kubenswrapper[4934]: I1227 07:57:18.490538 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81f8d777-8d08-49e0-b3c2-5ef2cb2388e3-utilities\") pod \"redhat-operators-wc8dz\" (UID: \"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3\") " pod="openshift-marketplace/redhat-operators-wc8dz" Dec 27 07:57:18 crc kubenswrapper[4934]: I1227 07:57:18.490581 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81f8d777-8d08-49e0-b3c2-5ef2cb2388e3-catalog-content\") pod \"redhat-operators-wc8dz\" (UID: \"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3\") " pod="openshift-marketplace/redhat-operators-wc8dz" Dec 27 07:57:18 crc kubenswrapper[4934]: I1227 07:57:18.491147 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81f8d777-8d08-49e0-b3c2-5ef2cb2388e3-utilities\") pod \"redhat-operators-wc8dz\" (UID: \"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3\") " pod="openshift-marketplace/redhat-operators-wc8dz" Dec 27 07:57:18 crc kubenswrapper[4934]: I1227 07:57:18.492036 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81f8d777-8d08-49e0-b3c2-5ef2cb2388e3-catalog-content\") pod \"redhat-operators-wc8dz\" (UID: \"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3\") " pod="openshift-marketplace/redhat-operators-wc8dz" Dec 27 07:57:18 crc kubenswrapper[4934]: I1227 07:57:18.510888 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw2v2\" (UniqueName: \"kubernetes.io/projected/81f8d777-8d08-49e0-b3c2-5ef2cb2388e3-kube-api-access-vw2v2\") pod \"redhat-operators-wc8dz\" (UID: \"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3\") " pod="openshift-marketplace/redhat-operators-wc8dz" Dec 27 07:57:18 crc kubenswrapper[4934]: I1227 07:57:18.592344 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wc8dz" Dec 27 07:57:19 crc kubenswrapper[4934]: I1227 07:57:19.001673 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wc8dz"] Dec 27 07:57:19 crc kubenswrapper[4934]: I1227 07:57:19.430641 4934 generic.go:334] "Generic (PLEG): container finished" podID="81f8d777-8d08-49e0-b3c2-5ef2cb2388e3" containerID="5047f910c09c6d5299ffb7b72d1b977f7eaa0b12205624d21d370b33b25afa10" exitCode=0 Dec 27 07:57:19 crc kubenswrapper[4934]: I1227 07:57:19.430750 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wc8dz" event={"ID":"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3","Type":"ContainerDied","Data":"5047f910c09c6d5299ffb7b72d1b977f7eaa0b12205624d21d370b33b25afa10"} Dec 27 07:57:19 crc kubenswrapper[4934]: I1227 07:57:19.431013 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wc8dz" event={"ID":"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3","Type":"ContainerStarted","Data":"55b397b78d86c357a757a2669a20680e08d8db047b074302dbb1957e9c211228"} Dec 27 07:57:19 crc kubenswrapper[4934]: I1227 07:57:19.433434 4934 generic.go:334] "Generic (PLEG): container finished" podID="7f133bd4-74ca-4b6c-bd9a-210102808e22" containerID="2ba2399727aa49b66d1f89af8842847661feefb8b2ff0220b398ca5324f840ac" exitCode=0 Dec 27 07:57:19 crc kubenswrapper[4934]: I1227 07:57:19.433468 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp" event={"ID":"7f133bd4-74ca-4b6c-bd9a-210102808e22","Type":"ContainerDied","Data":"2ba2399727aa49b66d1f89af8842847661feefb8b2ff0220b398ca5324f840ac"} Dec 27 07:57:20 crc kubenswrapper[4934]: I1227 07:57:20.444427 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wc8dz" event={"ID":"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3","Type":"ContainerStarted","Data":"b8bc9b2abd56f8a62bbb1ab78e868e06fb973bec16798cb2d325ae9b181cf070"} Dec 27 07:57:20 crc kubenswrapper[4934]: I1227 07:57:20.447897 4934 generic.go:334] "Generic (PLEG): container finished" podID="7f133bd4-74ca-4b6c-bd9a-210102808e22" containerID="83837ff6f9d71d09bab1d0b94b3aa10c4eb8449b4eed446ee307ffb1e7204dbb" exitCode=0 Dec 27 07:57:20 crc kubenswrapper[4934]: I1227 07:57:20.447991 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp" event={"ID":"7f133bd4-74ca-4b6c-bd9a-210102808e22","Type":"ContainerDied","Data":"83837ff6f9d71d09bab1d0b94b3aa10c4eb8449b4eed446ee307ffb1e7204dbb"} Dec 27 07:57:21 crc kubenswrapper[4934]: I1227 07:57:21.456048 4934 generic.go:334] "Generic (PLEG): container finished" podID="81f8d777-8d08-49e0-b3c2-5ef2cb2388e3" containerID="b8bc9b2abd56f8a62bbb1ab78e868e06fb973bec16798cb2d325ae9b181cf070" exitCode=0 Dec 27 07:57:21 crc kubenswrapper[4934]: I1227 07:57:21.456103 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wc8dz" event={"ID":"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3","Type":"ContainerDied","Data":"b8bc9b2abd56f8a62bbb1ab78e868e06fb973bec16798cb2d325ae9b181cf070"} Dec 27 07:57:21 crc kubenswrapper[4934]: I1227 07:57:21.847378 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp" Dec 27 07:57:21 crc kubenswrapper[4934]: I1227 07:57:21.948694 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7f133bd4-74ca-4b6c-bd9a-210102808e22-util\") pod \"7f133bd4-74ca-4b6c-bd9a-210102808e22\" (UID: \"7f133bd4-74ca-4b6c-bd9a-210102808e22\") " Dec 27 07:57:21 crc kubenswrapper[4934]: I1227 07:57:21.949068 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8sd2s\" (UniqueName: \"kubernetes.io/projected/7f133bd4-74ca-4b6c-bd9a-210102808e22-kube-api-access-8sd2s\") pod \"7f133bd4-74ca-4b6c-bd9a-210102808e22\" (UID: \"7f133bd4-74ca-4b6c-bd9a-210102808e22\") " Dec 27 07:57:21 crc kubenswrapper[4934]: I1227 07:57:21.949130 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7f133bd4-74ca-4b6c-bd9a-210102808e22-bundle\") pod \"7f133bd4-74ca-4b6c-bd9a-210102808e22\" (UID: \"7f133bd4-74ca-4b6c-bd9a-210102808e22\") " Dec 27 07:57:21 crc kubenswrapper[4934]: I1227 07:57:21.983645 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f133bd4-74ca-4b6c-bd9a-210102808e22-bundle" (OuterVolumeSpecName: "bundle") pod "7f133bd4-74ca-4b6c-bd9a-210102808e22" (UID: "7f133bd4-74ca-4b6c-bd9a-210102808e22"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:57:22 crc kubenswrapper[4934]: I1227 07:57:22.000450 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f133bd4-74ca-4b6c-bd9a-210102808e22-util" (OuterVolumeSpecName: "util") pod "7f133bd4-74ca-4b6c-bd9a-210102808e22" (UID: "7f133bd4-74ca-4b6c-bd9a-210102808e22"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:57:22 crc kubenswrapper[4934]: I1227 07:57:22.011523 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f133bd4-74ca-4b6c-bd9a-210102808e22-kube-api-access-8sd2s" (OuterVolumeSpecName: "kube-api-access-8sd2s") pod "7f133bd4-74ca-4b6c-bd9a-210102808e22" (UID: "7f133bd4-74ca-4b6c-bd9a-210102808e22"). InnerVolumeSpecName "kube-api-access-8sd2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:57:22 crc kubenswrapper[4934]: I1227 07:57:22.051328 4934 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7f133bd4-74ca-4b6c-bd9a-210102808e22-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 07:57:22 crc kubenswrapper[4934]: I1227 07:57:22.051361 4934 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7f133bd4-74ca-4b6c-bd9a-210102808e22-util\") on node \"crc\" DevicePath \"\"" Dec 27 07:57:22 crc kubenswrapper[4934]: I1227 07:57:22.051371 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8sd2s\" (UniqueName: \"kubernetes.io/projected/7f133bd4-74ca-4b6c-bd9a-210102808e22-kube-api-access-8sd2s\") on node \"crc\" DevicePath \"\"" Dec 27 07:57:22 crc kubenswrapper[4934]: I1227 07:57:22.470992 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wc8dz" event={"ID":"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3","Type":"ContainerStarted","Data":"930aad7437d81c315d49090929c9c81644f3b7db51796a9b5b28b37d0435e8e5"} Dec 27 07:57:22 crc kubenswrapper[4934]: I1227 07:57:22.474194 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp" event={"ID":"7f133bd4-74ca-4b6c-bd9a-210102808e22","Type":"ContainerDied","Data":"b53feb2edd2913969b2adde2698344d52d0010c2d6f5f8224229f66a83953376"} Dec 27 07:57:22 crc kubenswrapper[4934]: I1227 07:57:22.474240 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b53feb2edd2913969b2adde2698344d52d0010c2d6f5f8224229f66a83953376" Dec 27 07:57:22 crc kubenswrapper[4934]: I1227 07:57:22.474311 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp" Dec 27 07:57:22 crc kubenswrapper[4934]: I1227 07:57:22.493380 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wc8dz" podStartSLOduration=2.050202435 podStartE2EDuration="4.493362789s" podCreationTimestamp="2025-12-27 07:57:18 +0000 UTC" firstStartedPulling="2025-12-27 07:57:19.43198404 +0000 UTC m=+900.252424634" lastFinishedPulling="2025-12-27 07:57:21.875144394 +0000 UTC m=+902.695584988" observedRunningTime="2025-12-27 07:57:22.490275891 +0000 UTC m=+903.310716505" watchObservedRunningTime="2025-12-27 07:57:22.493362789 +0000 UTC m=+903.313803403" Dec 27 07:57:25 crc kubenswrapper[4934]: I1227 07:57:25.303304 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-p29kj"] Dec 27 07:57:25 crc kubenswrapper[4934]: E1227 07:57:25.304912 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f133bd4-74ca-4b6c-bd9a-210102808e22" containerName="pull" Dec 27 07:57:25 crc kubenswrapper[4934]: I1227 07:57:25.304960 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f133bd4-74ca-4b6c-bd9a-210102808e22" containerName="pull" Dec 27 07:57:25 crc kubenswrapper[4934]: E1227 07:57:25.304989 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f133bd4-74ca-4b6c-bd9a-210102808e22" containerName="extract" Dec 27 07:57:25 crc kubenswrapper[4934]: I1227 07:57:25.304998 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f133bd4-74ca-4b6c-bd9a-210102808e22" containerName="extract" Dec 27 07:57:25 crc kubenswrapper[4934]: E1227 07:57:25.305024 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f133bd4-74ca-4b6c-bd9a-210102808e22" containerName="util" Dec 27 07:57:25 crc kubenswrapper[4934]: I1227 07:57:25.305032 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f133bd4-74ca-4b6c-bd9a-210102808e22" containerName="util" Dec 27 07:57:25 crc kubenswrapper[4934]: I1227 07:57:25.305708 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f133bd4-74ca-4b6c-bd9a-210102808e22" containerName="extract" Dec 27 07:57:25 crc kubenswrapper[4934]: I1227 07:57:25.307366 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-6769fb99d-p29kj" Dec 27 07:57:25 crc kubenswrapper[4934]: I1227 07:57:25.317161 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-zjwq2" Dec 27 07:57:25 crc kubenswrapper[4934]: I1227 07:57:25.317461 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 27 07:57:25 crc kubenswrapper[4934]: I1227 07:57:25.317601 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 27 07:57:25 crc kubenswrapper[4934]: I1227 07:57:25.334152 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-p29kj"] Dec 27 07:57:25 crc kubenswrapper[4934]: I1227 07:57:25.402916 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf79p\" (UniqueName: \"kubernetes.io/projected/e7bfbaac-ef25-4807-8337-14656ff53f15-kube-api-access-wf79p\") pod \"nmstate-operator-6769fb99d-p29kj\" (UID: \"e7bfbaac-ef25-4807-8337-14656ff53f15\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-p29kj" Dec 27 07:57:25 crc kubenswrapper[4934]: I1227 07:57:25.504791 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf79p\" (UniqueName: \"kubernetes.io/projected/e7bfbaac-ef25-4807-8337-14656ff53f15-kube-api-access-wf79p\") pod \"nmstate-operator-6769fb99d-p29kj\" (UID: \"e7bfbaac-ef25-4807-8337-14656ff53f15\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-p29kj" Dec 27 07:57:25 crc kubenswrapper[4934]: I1227 07:57:25.521953 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf79p\" (UniqueName: \"kubernetes.io/projected/e7bfbaac-ef25-4807-8337-14656ff53f15-kube-api-access-wf79p\") pod \"nmstate-operator-6769fb99d-p29kj\" (UID: \"e7bfbaac-ef25-4807-8337-14656ff53f15\") " pod="openshift-nmstate/nmstate-operator-6769fb99d-p29kj" Dec 27 07:57:25 crc kubenswrapper[4934]: I1227 07:57:25.643840 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-6769fb99d-p29kj" Dec 27 07:57:26 crc kubenswrapper[4934]: I1227 07:57:26.159492 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-6769fb99d-p29kj"] Dec 27 07:57:26 crc kubenswrapper[4934]: W1227 07:57:26.165337 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7bfbaac_ef25_4807_8337_14656ff53f15.slice/crio-febbb5a6147f89f6814ec6ee005800e5a19f4b36067ce7d70dbe23c28b89570a WatchSource:0}: Error finding container febbb5a6147f89f6814ec6ee005800e5a19f4b36067ce7d70dbe23c28b89570a: Status 404 returned error can't find the container with id febbb5a6147f89f6814ec6ee005800e5a19f4b36067ce7d70dbe23c28b89570a Dec 27 07:57:26 crc kubenswrapper[4934]: I1227 07:57:26.507201 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-6769fb99d-p29kj" event={"ID":"e7bfbaac-ef25-4807-8337-14656ff53f15","Type":"ContainerStarted","Data":"febbb5a6147f89f6814ec6ee005800e5a19f4b36067ce7d70dbe23c28b89570a"} Dec 27 07:57:28 crc kubenswrapper[4934]: I1227 07:57:28.592651 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wc8dz" Dec 27 07:57:28 crc kubenswrapper[4934]: I1227 07:57:28.592981 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wc8dz" Dec 27 07:57:28 crc kubenswrapper[4934]: I1227 07:57:28.631433 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wc8dz" Dec 27 07:57:29 crc kubenswrapper[4934]: I1227 07:57:29.570620 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wc8dz" Dec 27 07:57:30 crc kubenswrapper[4934]: I1227 07:57:30.538015 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-6769fb99d-p29kj" event={"ID":"e7bfbaac-ef25-4807-8337-14656ff53f15","Type":"ContainerStarted","Data":"98a3b60dedc50e3040b2d6c83f80786e54cb691fe0ebc141cd4aed9b047de6a5"} Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.054953 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-6769fb99d-p29kj" podStartSLOduration=2.491315758 podStartE2EDuration="6.054935082s" podCreationTimestamp="2025-12-27 07:57:25 +0000 UTC" firstStartedPulling="2025-12-27 07:57:26.167914207 +0000 UTC m=+906.988354811" lastFinishedPulling="2025-12-27 07:57:29.731533501 +0000 UTC m=+910.551974135" observedRunningTime="2025-12-27 07:57:30.559666653 +0000 UTC m=+911.380107247" watchObservedRunningTime="2025-12-27 07:57:31.054935082 +0000 UTC m=+911.875375686" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.060565 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wc8dz"] Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.487287 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-6fp5z"] Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.488681 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-6fp5z" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.490294 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-xpgpn" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.502376 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-6fp5z"] Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.516313 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt"] Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.517551 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.521602 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.535261 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-kbxfg"] Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.536180 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-kbxfg" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.542848 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wc8dz" podUID="81f8d777-8d08-49e0-b3c2-5ef2cb2388e3" containerName="registry-server" containerID="cri-o://930aad7437d81c315d49090929c9c81644f3b7db51796a9b5b28b37d0435e8e5" gracePeriod=2 Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.549318 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt"] Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.605526 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k25ql\" (UniqueName: \"kubernetes.io/projected/4f0bb073-052f-45eb-9cd6-020b9323b2ca-kube-api-access-k25ql\") pod \"nmstate-metrics-7f7f7578db-6fp5z\" (UID: \"4f0bb073-052f-45eb-9cd6-020b9323b2ca\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-6fp5z" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.690666 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-7hbj6"] Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.692064 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-7hbj6" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.694670 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-2kdsn" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.695118 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.695434 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.697772 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-7hbj6"] Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.707623 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/415e9aa6-5a7a-4d93-924d-0213c9a5ca4d-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-fwjdt\" (UID: \"415e9aa6-5a7a-4d93-924d-0213c9a5ca4d\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.707691 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9krr\" (UniqueName: \"kubernetes.io/projected/415e9aa6-5a7a-4d93-924d-0213c9a5ca4d-kube-api-access-r9krr\") pod \"nmstate-webhook-f8fb84555-fwjdt\" (UID: \"415e9aa6-5a7a-4d93-924d-0213c9a5ca4d\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.707793 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/224f9a57-4665-4c0f-865c-ac2267f69b7e-nmstate-lock\") pod \"nmstate-handler-kbxfg\" (UID: \"224f9a57-4665-4c0f-865c-ac2267f69b7e\") " pod="openshift-nmstate/nmstate-handler-kbxfg" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.707833 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k25ql\" (UniqueName: \"kubernetes.io/projected/4f0bb073-052f-45eb-9cd6-020b9323b2ca-kube-api-access-k25ql\") pod \"nmstate-metrics-7f7f7578db-6fp5z\" (UID: \"4f0bb073-052f-45eb-9cd6-020b9323b2ca\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-6fp5z" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.707861 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw5vv\" (UniqueName: \"kubernetes.io/projected/224f9a57-4665-4c0f-865c-ac2267f69b7e-kube-api-access-kw5vv\") pod \"nmstate-handler-kbxfg\" (UID: \"224f9a57-4665-4c0f-865c-ac2267f69b7e\") " pod="openshift-nmstate/nmstate-handler-kbxfg" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.707933 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/224f9a57-4665-4c0f-865c-ac2267f69b7e-dbus-socket\") pod \"nmstate-handler-kbxfg\" (UID: \"224f9a57-4665-4c0f-865c-ac2267f69b7e\") " pod="openshift-nmstate/nmstate-handler-kbxfg" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.707965 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/224f9a57-4665-4c0f-865c-ac2267f69b7e-ovs-socket\") pod \"nmstate-handler-kbxfg\" (UID: \"224f9a57-4665-4c0f-865c-ac2267f69b7e\") " pod="openshift-nmstate/nmstate-handler-kbxfg" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.744563 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k25ql\" (UniqueName: \"kubernetes.io/projected/4f0bb073-052f-45eb-9cd6-020b9323b2ca-kube-api-access-k25ql\") pod \"nmstate-metrics-7f7f7578db-6fp5z\" (UID: \"4f0bb073-052f-45eb-9cd6-020b9323b2ca\") " pod="openshift-nmstate/nmstate-metrics-7f7f7578db-6fp5z" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.804035 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-6fp5z" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.812066 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vhxm\" (UniqueName: \"kubernetes.io/projected/60914d81-fa54-4836-8b55-2cbe909df91d-kube-api-access-8vhxm\") pod \"nmstate-console-plugin-6ff7998486-7hbj6\" (UID: \"60914d81-fa54-4836-8b55-2cbe909df91d\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-7hbj6" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.812146 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/415e9aa6-5a7a-4d93-924d-0213c9a5ca4d-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-fwjdt\" (UID: \"415e9aa6-5a7a-4d93-924d-0213c9a5ca4d\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.812180 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9krr\" (UniqueName: \"kubernetes.io/projected/415e9aa6-5a7a-4d93-924d-0213c9a5ca4d-kube-api-access-r9krr\") pod \"nmstate-webhook-f8fb84555-fwjdt\" (UID: \"415e9aa6-5a7a-4d93-924d-0213c9a5ca4d\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.812202 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/60914d81-fa54-4836-8b55-2cbe909df91d-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-7hbj6\" (UID: \"60914d81-fa54-4836-8b55-2cbe909df91d\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-7hbj6" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.812248 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/60914d81-fa54-4836-8b55-2cbe909df91d-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-7hbj6\" (UID: \"60914d81-fa54-4836-8b55-2cbe909df91d\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-7hbj6" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.812282 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/224f9a57-4665-4c0f-865c-ac2267f69b7e-nmstate-lock\") pod \"nmstate-handler-kbxfg\" (UID: \"224f9a57-4665-4c0f-865c-ac2267f69b7e\") " pod="openshift-nmstate/nmstate-handler-kbxfg" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.812312 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw5vv\" (UniqueName: \"kubernetes.io/projected/224f9a57-4665-4c0f-865c-ac2267f69b7e-kube-api-access-kw5vv\") pod \"nmstate-handler-kbxfg\" (UID: \"224f9a57-4665-4c0f-865c-ac2267f69b7e\") " pod="openshift-nmstate/nmstate-handler-kbxfg" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.812337 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/224f9a57-4665-4c0f-865c-ac2267f69b7e-dbus-socket\") pod \"nmstate-handler-kbxfg\" (UID: \"224f9a57-4665-4c0f-865c-ac2267f69b7e\") " pod="openshift-nmstate/nmstate-handler-kbxfg" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.812359 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/224f9a57-4665-4c0f-865c-ac2267f69b7e-ovs-socket\") pod \"nmstate-handler-kbxfg\" (UID: \"224f9a57-4665-4c0f-865c-ac2267f69b7e\") " pod="openshift-nmstate/nmstate-handler-kbxfg" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.812433 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/224f9a57-4665-4c0f-865c-ac2267f69b7e-ovs-socket\") pod \"nmstate-handler-kbxfg\" (UID: \"224f9a57-4665-4c0f-865c-ac2267f69b7e\") " pod="openshift-nmstate/nmstate-handler-kbxfg" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.816727 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/415e9aa6-5a7a-4d93-924d-0213c9a5ca4d-tls-key-pair\") pod \"nmstate-webhook-f8fb84555-fwjdt\" (UID: \"415e9aa6-5a7a-4d93-924d-0213c9a5ca4d\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.816775 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/224f9a57-4665-4c0f-865c-ac2267f69b7e-nmstate-lock\") pod \"nmstate-handler-kbxfg\" (UID: \"224f9a57-4665-4c0f-865c-ac2267f69b7e\") " pod="openshift-nmstate/nmstate-handler-kbxfg" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.816898 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/224f9a57-4665-4c0f-865c-ac2267f69b7e-dbus-socket\") pod \"nmstate-handler-kbxfg\" (UID: \"224f9a57-4665-4c0f-865c-ac2267f69b7e\") " pod="openshift-nmstate/nmstate-handler-kbxfg" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.831969 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9krr\" (UniqueName: \"kubernetes.io/projected/415e9aa6-5a7a-4d93-924d-0213c9a5ca4d-kube-api-access-r9krr\") pod \"nmstate-webhook-f8fb84555-fwjdt\" (UID: \"415e9aa6-5a7a-4d93-924d-0213c9a5ca4d\") " pod="openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.835151 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.846841 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw5vv\" (UniqueName: \"kubernetes.io/projected/224f9a57-4665-4c0f-865c-ac2267f69b7e-kube-api-access-kw5vv\") pod \"nmstate-handler-kbxfg\" (UID: \"224f9a57-4665-4c0f-865c-ac2267f69b7e\") " pod="openshift-nmstate/nmstate-handler-kbxfg" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.851407 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-kbxfg" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.874924 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-86b5d5c74c-c9lp9"] Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.875769 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.891005 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-86b5d5c74c-c9lp9"] Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.936271 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vhxm\" (UniqueName: \"kubernetes.io/projected/60914d81-fa54-4836-8b55-2cbe909df91d-kube-api-access-8vhxm\") pod \"nmstate-console-plugin-6ff7998486-7hbj6\" (UID: \"60914d81-fa54-4836-8b55-2cbe909df91d\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-7hbj6" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.936408 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/60914d81-fa54-4836-8b55-2cbe909df91d-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-7hbj6\" (UID: \"60914d81-fa54-4836-8b55-2cbe909df91d\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-7hbj6" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.936545 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/60914d81-fa54-4836-8b55-2cbe909df91d-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-7hbj6\" (UID: \"60914d81-fa54-4836-8b55-2cbe909df91d\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-7hbj6" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.940881 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/60914d81-fa54-4836-8b55-2cbe909df91d-nginx-conf\") pod \"nmstate-console-plugin-6ff7998486-7hbj6\" (UID: \"60914d81-fa54-4836-8b55-2cbe909df91d\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-7hbj6" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.959183 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vhxm\" (UniqueName: \"kubernetes.io/projected/60914d81-fa54-4836-8b55-2cbe909df91d-kube-api-access-8vhxm\") pod \"nmstate-console-plugin-6ff7998486-7hbj6\" (UID: \"60914d81-fa54-4836-8b55-2cbe909df91d\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-7hbj6" Dec 27 07:57:31 crc kubenswrapper[4934]: I1227 07:57:31.960850 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/60914d81-fa54-4836-8b55-2cbe909df91d-plugin-serving-cert\") pod \"nmstate-console-plugin-6ff7998486-7hbj6\" (UID: \"60914d81-fa54-4836-8b55-2cbe909df91d\") " pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-7hbj6" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.015822 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-7hbj6" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.042504 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-trusted-ca-bundle\") pod \"console-86b5d5c74c-c9lp9\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.042761 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-oauth-serving-cert\") pod \"console-86b5d5c74c-c9lp9\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.042788 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj5rv\" (UniqueName: \"kubernetes.io/projected/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-kube-api-access-cj5rv\") pod \"console-86b5d5c74c-c9lp9\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.042805 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-console-serving-cert\") pod \"console-86b5d5c74c-c9lp9\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.042844 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-console-config\") pod \"console-86b5d5c74c-c9lp9\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.043003 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-console-oauth-config\") pod \"console-86b5d5c74c-c9lp9\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.043206 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-service-ca\") pod \"console-86b5d5c74c-c9lp9\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.144609 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-console-config\") pod \"console-86b5d5c74c-c9lp9\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.144669 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-console-oauth-config\") pod \"console-86b5d5c74c-c9lp9\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.144706 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-service-ca\") pod \"console-86b5d5c74c-c9lp9\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.144735 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-trusted-ca-bundle\") pod \"console-86b5d5c74c-c9lp9\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.144790 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-oauth-serving-cert\") pod \"console-86b5d5c74c-c9lp9\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.144821 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj5rv\" (UniqueName: \"kubernetes.io/projected/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-kube-api-access-cj5rv\") pod \"console-86b5d5c74c-c9lp9\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.144841 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-console-serving-cert\") pod \"console-86b5d5c74c-c9lp9\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.146169 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-service-ca\") pod \"console-86b5d5c74c-c9lp9\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.146419 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-console-config\") pod \"console-86b5d5c74c-c9lp9\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.147263 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-oauth-serving-cert\") pod \"console-86b5d5c74c-c9lp9\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.147323 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-trusted-ca-bundle\") pod \"console-86b5d5c74c-c9lp9\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.149573 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-console-serving-cert\") pod \"console-86b5d5c74c-c9lp9\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.150549 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-console-oauth-config\") pod \"console-86b5d5c74c-c9lp9\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.170581 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj5rv\" (UniqueName: \"kubernetes.io/projected/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-kube-api-access-cj5rv\") pod \"console-86b5d5c74c-c9lp9\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.205576 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt"] Dec 27 07:57:32 crc kubenswrapper[4934]: W1227 07:57:32.214312 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod415e9aa6_5a7a_4d93_924d_0213c9a5ca4d.slice/crio-df8bec0ea5e897a0022628a1aad84373390a72b23f0b7fe89c4cbccd65b5a912 WatchSource:0}: Error finding container df8bec0ea5e897a0022628a1aad84373390a72b23f0b7fe89c4cbccd65b5a912: Status 404 returned error can't find the container with id df8bec0ea5e897a0022628a1aad84373390a72b23f0b7fe89c4cbccd65b5a912 Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.258781 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.466185 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f7f7578db-6fp5z"] Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.550986 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6ff7998486-7hbj6"] Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.552461 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-6fp5z" event={"ID":"4f0bb073-052f-45eb-9cd6-020b9323b2ca","Type":"ContainerStarted","Data":"7ff8b25e2fa3af651fada522078dd178c2fb3cc22238c1be143d5d765036cb16"} Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.554103 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt" event={"ID":"415e9aa6-5a7a-4d93-924d-0213c9a5ca4d","Type":"ContainerStarted","Data":"df8bec0ea5e897a0022628a1aad84373390a72b23f0b7fe89c4cbccd65b5a912"} Dec 27 07:57:32 crc kubenswrapper[4934]: W1227 07:57:32.554466 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60914d81_fa54_4836_8b55_2cbe909df91d.slice/crio-04d3be8cc5ea04eb44b23276e4edd9781f041097678f64795881d3c67b515904 WatchSource:0}: Error finding container 04d3be8cc5ea04eb44b23276e4edd9781f041097678f64795881d3c67b515904: Status 404 returned error can't find the container with id 04d3be8cc5ea04eb44b23276e4edd9781f041097678f64795881d3c67b515904 Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.555837 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-kbxfg" event={"ID":"224f9a57-4665-4c0f-865c-ac2267f69b7e","Type":"ContainerStarted","Data":"778467ea82a09766333d566fe936a002c9f748295d3cc387212dfdc97083af74"} Dec 27 07:57:32 crc kubenswrapper[4934]: W1227 07:57:32.700768 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8aa711d1_ae51_480c_aa8b_9a510dfd7b6a.slice/crio-32a5fad3b33b80f1a9941ce5b483e68c94c17a6b404184f3df8a23d351fecc9b WatchSource:0}: Error finding container 32a5fad3b33b80f1a9941ce5b483e68c94c17a6b404184f3df8a23d351fecc9b: Status 404 returned error can't find the container with id 32a5fad3b33b80f1a9941ce5b483e68c94c17a6b404184f3df8a23d351fecc9b Dec 27 07:57:32 crc kubenswrapper[4934]: I1227 07:57:32.702014 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-86b5d5c74c-c9lp9"] Dec 27 07:57:33 crc kubenswrapper[4934]: I1227 07:57:33.564888 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-7hbj6" event={"ID":"60914d81-fa54-4836-8b55-2cbe909df91d","Type":"ContainerStarted","Data":"04d3be8cc5ea04eb44b23276e4edd9781f041097678f64795881d3c67b515904"} Dec 27 07:57:33 crc kubenswrapper[4934]: I1227 07:57:33.567793 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-86b5d5c74c-c9lp9" event={"ID":"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a","Type":"ContainerStarted","Data":"32a5fad3b33b80f1a9941ce5b483e68c94c17a6b404184f3df8a23d351fecc9b"} Dec 27 07:57:34 crc kubenswrapper[4934]: I1227 07:57:34.579295 4934 generic.go:334] "Generic (PLEG): container finished" podID="81f8d777-8d08-49e0-b3c2-5ef2cb2388e3" containerID="930aad7437d81c315d49090929c9c81644f3b7db51796a9b5b28b37d0435e8e5" exitCode=0 Dec 27 07:57:34 crc kubenswrapper[4934]: I1227 07:57:34.579362 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wc8dz" event={"ID":"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3","Type":"ContainerDied","Data":"930aad7437d81c315d49090929c9c81644f3b7db51796a9b5b28b37d0435e8e5"} Dec 27 07:57:34 crc kubenswrapper[4934]: I1227 07:57:34.582269 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-86b5d5c74c-c9lp9" event={"ID":"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a","Type":"ContainerStarted","Data":"eacd15e000e9c4e9be81a96aad383398017a9643d62d695c37cfa5fb8b300d93"} Dec 27 07:57:34 crc kubenswrapper[4934]: I1227 07:57:34.604021 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-86b5d5c74c-c9lp9" podStartSLOduration=3.604000728 podStartE2EDuration="3.604000728s" podCreationTimestamp="2025-12-27 07:57:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:57:34.598176761 +0000 UTC m=+915.418617365" watchObservedRunningTime="2025-12-27 07:57:34.604000728 +0000 UTC m=+915.424441322" Dec 27 07:57:34 crc kubenswrapper[4934]: I1227 07:57:34.908498 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wc8dz" Dec 27 07:57:34 crc kubenswrapper[4934]: I1227 07:57:34.992296 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81f8d777-8d08-49e0-b3c2-5ef2cb2388e3-catalog-content\") pod \"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3\" (UID: \"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3\") " Dec 27 07:57:34 crc kubenswrapper[4934]: I1227 07:57:34.992466 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vw2v2\" (UniqueName: \"kubernetes.io/projected/81f8d777-8d08-49e0-b3c2-5ef2cb2388e3-kube-api-access-vw2v2\") pod \"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3\" (UID: \"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3\") " Dec 27 07:57:34 crc kubenswrapper[4934]: I1227 07:57:34.992494 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81f8d777-8d08-49e0-b3c2-5ef2cb2388e3-utilities\") pod \"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3\" (UID: \"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3\") " Dec 27 07:57:34 crc kubenswrapper[4934]: I1227 07:57:34.993501 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81f8d777-8d08-49e0-b3c2-5ef2cb2388e3-utilities" (OuterVolumeSpecName: "utilities") pod "81f8d777-8d08-49e0-b3c2-5ef2cb2388e3" (UID: "81f8d777-8d08-49e0-b3c2-5ef2cb2388e3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:57:34 crc kubenswrapper[4934]: I1227 07:57:34.994151 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81f8d777-8d08-49e0-b3c2-5ef2cb2388e3-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 07:57:35 crc kubenswrapper[4934]: I1227 07:57:35.002423 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81f8d777-8d08-49e0-b3c2-5ef2cb2388e3-kube-api-access-vw2v2" (OuterVolumeSpecName: "kube-api-access-vw2v2") pod "81f8d777-8d08-49e0-b3c2-5ef2cb2388e3" (UID: "81f8d777-8d08-49e0-b3c2-5ef2cb2388e3"). InnerVolumeSpecName "kube-api-access-vw2v2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:57:35 crc kubenswrapper[4934]: I1227 07:57:35.105380 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vw2v2\" (UniqueName: \"kubernetes.io/projected/81f8d777-8d08-49e0-b3c2-5ef2cb2388e3-kube-api-access-vw2v2\") on node \"crc\" DevicePath \"\"" Dec 27 07:57:35 crc kubenswrapper[4934]: I1227 07:57:35.113709 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81f8d777-8d08-49e0-b3c2-5ef2cb2388e3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "81f8d777-8d08-49e0-b3c2-5ef2cb2388e3" (UID: "81f8d777-8d08-49e0-b3c2-5ef2cb2388e3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:57:35 crc kubenswrapper[4934]: I1227 07:57:35.207068 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81f8d777-8d08-49e0-b3c2-5ef2cb2388e3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 07:57:35 crc kubenswrapper[4934]: I1227 07:57:35.605372 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wc8dz" Dec 27 07:57:35 crc kubenswrapper[4934]: I1227 07:57:35.606825 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wc8dz" event={"ID":"81f8d777-8d08-49e0-b3c2-5ef2cb2388e3","Type":"ContainerDied","Data":"55b397b78d86c357a757a2669a20680e08d8db047b074302dbb1957e9c211228"} Dec 27 07:57:35 crc kubenswrapper[4934]: I1227 07:57:35.606898 4934 scope.go:117] "RemoveContainer" containerID="930aad7437d81c315d49090929c9c81644f3b7db51796a9b5b28b37d0435e8e5" Dec 27 07:57:35 crc kubenswrapper[4934]: I1227 07:57:35.631738 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wc8dz"] Dec 27 07:57:35 crc kubenswrapper[4934]: I1227 07:57:35.643027 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wc8dz"] Dec 27 07:57:35 crc kubenswrapper[4934]: I1227 07:57:35.760504 4934 scope.go:117] "RemoveContainer" containerID="b8bc9b2abd56f8a62bbb1ab78e868e06fb973bec16798cb2d325ae9b181cf070" Dec 27 07:57:35 crc kubenswrapper[4934]: I1227 07:57:35.795478 4934 scope.go:117] "RemoveContainer" containerID="5047f910c09c6d5299ffb7b72d1b977f7eaa0b12205624d21d370b33b25afa10" Dec 27 07:57:36 crc kubenswrapper[4934]: I1227 07:57:36.613706 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-6fp5z" event={"ID":"4f0bb073-052f-45eb-9cd6-020b9323b2ca","Type":"ContainerStarted","Data":"e98433280619adf9332e9c907d29b3d8f8fc72a85fd8e5e3eb4265e198eec497"} Dec 27 07:57:36 crc kubenswrapper[4934]: I1227 07:57:36.620227 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt" event={"ID":"415e9aa6-5a7a-4d93-924d-0213c9a5ca4d","Type":"ContainerStarted","Data":"2ddb845251e88350c95fd543d6f7a8135c22cd3a1e61ddf319a256b4f4023538"} Dec 27 07:57:36 crc kubenswrapper[4934]: I1227 07:57:36.620315 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt" Dec 27 07:57:36 crc kubenswrapper[4934]: I1227 07:57:36.621938 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-kbxfg" event={"ID":"224f9a57-4665-4c0f-865c-ac2267f69b7e","Type":"ContainerStarted","Data":"18ce56ccd17df0a94ca33f724c0fbe2e3451a11660e6bf09eb6d39943b3b3b2a"} Dec 27 07:57:36 crc kubenswrapper[4934]: I1227 07:57:36.622324 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-kbxfg" Dec 27 07:57:36 crc kubenswrapper[4934]: I1227 07:57:36.623792 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-7hbj6" event={"ID":"60914d81-fa54-4836-8b55-2cbe909df91d","Type":"ContainerStarted","Data":"1848eb6f047b7b47b349b7194914ffe01d4cccce53b1e2ede48e29c7f43d9b2f"} Dec 27 07:57:36 crc kubenswrapper[4934]: I1227 07:57:36.643580 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt" podStartSLOduration=2.043450892 podStartE2EDuration="5.643557099s" podCreationTimestamp="2025-12-27 07:57:31 +0000 UTC" firstStartedPulling="2025-12-27 07:57:32.216532342 +0000 UTC m=+913.036972936" lastFinishedPulling="2025-12-27 07:57:35.816638529 +0000 UTC m=+916.637079143" observedRunningTime="2025-12-27 07:57:36.634971162 +0000 UTC m=+917.455411776" watchObservedRunningTime="2025-12-27 07:57:36.643557099 +0000 UTC m=+917.463997703" Dec 27 07:57:36 crc kubenswrapper[4934]: I1227 07:57:36.664606 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-kbxfg" podStartSLOduration=1.739855469 podStartE2EDuration="5.664578s" podCreationTimestamp="2025-12-27 07:57:31 +0000 UTC" firstStartedPulling="2025-12-27 07:57:31.906491356 +0000 UTC m=+912.726931950" lastFinishedPulling="2025-12-27 07:57:35.831213887 +0000 UTC m=+916.651654481" observedRunningTime="2025-12-27 07:57:36.658393364 +0000 UTC m=+917.478833998" watchObservedRunningTime="2025-12-27 07:57:36.664578 +0000 UTC m=+917.485018634" Dec 27 07:57:36 crc kubenswrapper[4934]: I1227 07:57:36.689497 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6ff7998486-7hbj6" podStartSLOduration=2.434418086 podStartE2EDuration="5.68947957s" podCreationTimestamp="2025-12-27 07:57:31 +0000 UTC" firstStartedPulling="2025-12-27 07:57:32.556300991 +0000 UTC m=+913.376741585" lastFinishedPulling="2025-12-27 07:57:35.811362465 +0000 UTC m=+916.631803069" observedRunningTime="2025-12-27 07:57:36.675767213 +0000 UTC m=+917.496207817" watchObservedRunningTime="2025-12-27 07:57:36.68947957 +0000 UTC m=+917.509920164" Dec 27 07:57:37 crc kubenswrapper[4934]: I1227 07:57:37.480289 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81f8d777-8d08-49e0-b3c2-5ef2cb2388e3" path="/var/lib/kubelet/pods/81f8d777-8d08-49e0-b3c2-5ef2cb2388e3/volumes" Dec 27 07:57:39 crc kubenswrapper[4934]: I1227 07:57:39.649658 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-6fp5z" event={"ID":"4f0bb073-052f-45eb-9cd6-020b9323b2ca","Type":"ContainerStarted","Data":"bc57012f176f306e58062691e81942e9fa3fe7ccc50bd2c2d2fa68fb05840c00"} Dec 27 07:57:39 crc kubenswrapper[4934]: I1227 07:57:39.675517 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f7f7578db-6fp5z" podStartSLOduration=2.493843578 podStartE2EDuration="8.675495165s" podCreationTimestamp="2025-12-27 07:57:31 +0000 UTC" firstStartedPulling="2025-12-27 07:57:32.467173418 +0000 UTC m=+913.287614012" lastFinishedPulling="2025-12-27 07:57:38.648824995 +0000 UTC m=+919.469265599" observedRunningTime="2025-12-27 07:57:39.664159289 +0000 UTC m=+920.484599873" watchObservedRunningTime="2025-12-27 07:57:39.675495165 +0000 UTC m=+920.495935759" Dec 27 07:57:41 crc kubenswrapper[4934]: I1227 07:57:41.875044 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-kbxfg" Dec 27 07:57:42 crc kubenswrapper[4934]: I1227 07:57:42.259583 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:42 crc kubenswrapper[4934]: I1227 07:57:42.259665 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:42 crc kubenswrapper[4934]: I1227 07:57:42.274495 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:42 crc kubenswrapper[4934]: I1227 07:57:42.672100 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 07:57:42 crc kubenswrapper[4934]: I1227 07:57:42.722188 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-5d9c4bb489-5mh5q"] Dec 27 07:57:45 crc kubenswrapper[4934]: I1227 07:57:45.330369 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 07:57:45 crc kubenswrapper[4934]: I1227 07:57:45.330692 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 07:57:51 crc kubenswrapper[4934]: I1227 07:57:51.843279 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt" Dec 27 07:57:53 crc kubenswrapper[4934]: I1227 07:57:53.220525 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wmzfj"] Dec 27 07:57:53 crc kubenswrapper[4934]: E1227 07:57:53.221239 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f8d777-8d08-49e0-b3c2-5ef2cb2388e3" containerName="extract-content" Dec 27 07:57:53 crc kubenswrapper[4934]: I1227 07:57:53.221259 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f8d777-8d08-49e0-b3c2-5ef2cb2388e3" containerName="extract-content" Dec 27 07:57:53 crc kubenswrapper[4934]: E1227 07:57:53.221287 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f8d777-8d08-49e0-b3c2-5ef2cb2388e3" containerName="registry-server" Dec 27 07:57:53 crc kubenswrapper[4934]: I1227 07:57:53.221300 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f8d777-8d08-49e0-b3c2-5ef2cb2388e3" containerName="registry-server" Dec 27 07:57:53 crc kubenswrapper[4934]: E1227 07:57:53.221329 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81f8d777-8d08-49e0-b3c2-5ef2cb2388e3" containerName="extract-utilities" Dec 27 07:57:53 crc kubenswrapper[4934]: I1227 07:57:53.221341 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="81f8d777-8d08-49e0-b3c2-5ef2cb2388e3" containerName="extract-utilities" Dec 27 07:57:53 crc kubenswrapper[4934]: I1227 07:57:53.221619 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="81f8d777-8d08-49e0-b3c2-5ef2cb2388e3" containerName="registry-server" Dec 27 07:57:53 crc kubenswrapper[4934]: I1227 07:57:53.223508 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wmzfj" Dec 27 07:57:53 crc kubenswrapper[4934]: I1227 07:57:53.228285 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wmzfj"] Dec 27 07:57:53 crc kubenswrapper[4934]: I1227 07:57:53.306174 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/412e15a1-b444-4187-a639-a171ff9a2732-catalog-content\") pod \"community-operators-wmzfj\" (UID: \"412e15a1-b444-4187-a639-a171ff9a2732\") " pod="openshift-marketplace/community-operators-wmzfj" Dec 27 07:57:53 crc kubenswrapper[4934]: I1227 07:57:53.306249 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9d8h\" (UniqueName: \"kubernetes.io/projected/412e15a1-b444-4187-a639-a171ff9a2732-kube-api-access-j9d8h\") pod \"community-operators-wmzfj\" (UID: \"412e15a1-b444-4187-a639-a171ff9a2732\") " pod="openshift-marketplace/community-operators-wmzfj" Dec 27 07:57:53 crc kubenswrapper[4934]: I1227 07:57:53.306292 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/412e15a1-b444-4187-a639-a171ff9a2732-utilities\") pod \"community-operators-wmzfj\" (UID: \"412e15a1-b444-4187-a639-a171ff9a2732\") " pod="openshift-marketplace/community-operators-wmzfj" Dec 27 07:57:53 crc kubenswrapper[4934]: I1227 07:57:53.407658 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9d8h\" (UniqueName: \"kubernetes.io/projected/412e15a1-b444-4187-a639-a171ff9a2732-kube-api-access-j9d8h\") pod \"community-operators-wmzfj\" (UID: \"412e15a1-b444-4187-a639-a171ff9a2732\") " pod="openshift-marketplace/community-operators-wmzfj" Dec 27 07:57:53 crc kubenswrapper[4934]: I1227 07:57:53.407729 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/412e15a1-b444-4187-a639-a171ff9a2732-utilities\") pod \"community-operators-wmzfj\" (UID: \"412e15a1-b444-4187-a639-a171ff9a2732\") " pod="openshift-marketplace/community-operators-wmzfj" Dec 27 07:57:53 crc kubenswrapper[4934]: I1227 07:57:53.407887 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/412e15a1-b444-4187-a639-a171ff9a2732-catalog-content\") pod \"community-operators-wmzfj\" (UID: \"412e15a1-b444-4187-a639-a171ff9a2732\") " pod="openshift-marketplace/community-operators-wmzfj" Dec 27 07:57:53 crc kubenswrapper[4934]: I1227 07:57:53.408473 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/412e15a1-b444-4187-a639-a171ff9a2732-utilities\") pod \"community-operators-wmzfj\" (UID: \"412e15a1-b444-4187-a639-a171ff9a2732\") " pod="openshift-marketplace/community-operators-wmzfj" Dec 27 07:57:53 crc kubenswrapper[4934]: I1227 07:57:53.408491 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/412e15a1-b444-4187-a639-a171ff9a2732-catalog-content\") pod \"community-operators-wmzfj\" (UID: \"412e15a1-b444-4187-a639-a171ff9a2732\") " pod="openshift-marketplace/community-operators-wmzfj" Dec 27 07:57:53 crc kubenswrapper[4934]: I1227 07:57:53.425553 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9d8h\" (UniqueName: \"kubernetes.io/projected/412e15a1-b444-4187-a639-a171ff9a2732-kube-api-access-j9d8h\") pod \"community-operators-wmzfj\" (UID: \"412e15a1-b444-4187-a639-a171ff9a2732\") " pod="openshift-marketplace/community-operators-wmzfj" Dec 27 07:57:53 crc kubenswrapper[4934]: I1227 07:57:53.552379 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wmzfj" Dec 27 07:57:53 crc kubenswrapper[4934]: I1227 07:57:53.882781 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wmzfj"] Dec 27 07:57:54 crc kubenswrapper[4934]: I1227 07:57:54.772181 4934 generic.go:334] "Generic (PLEG): container finished" podID="412e15a1-b444-4187-a639-a171ff9a2732" containerID="82c81277286d3969b89b1ad968d65e0858aed21f718abd35f37b3e820efe3c60" exitCode=0 Dec 27 07:57:54 crc kubenswrapper[4934]: I1227 07:57:54.772281 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wmzfj" event={"ID":"412e15a1-b444-4187-a639-a171ff9a2732","Type":"ContainerDied","Data":"82c81277286d3969b89b1ad968d65e0858aed21f718abd35f37b3e820efe3c60"} Dec 27 07:57:54 crc kubenswrapper[4934]: I1227 07:57:54.772625 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wmzfj" event={"ID":"412e15a1-b444-4187-a639-a171ff9a2732","Type":"ContainerStarted","Data":"d4531431cc5816f9713360c9c2f8251f9321e6fdd9720fab35d0b59580c28fdc"} Dec 27 07:57:54 crc kubenswrapper[4934]: I1227 07:57:54.774145 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 27 07:57:55 crc kubenswrapper[4934]: I1227 07:57:55.780361 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wmzfj" event={"ID":"412e15a1-b444-4187-a639-a171ff9a2732","Type":"ContainerStarted","Data":"df6bde3b4f1194874df74ce8620448010ffad5b9f135d449273bdad71a069f43"} Dec 27 07:57:56 crc kubenswrapper[4934]: I1227 07:57:56.797948 4934 generic.go:334] "Generic (PLEG): container finished" podID="412e15a1-b444-4187-a639-a171ff9a2732" containerID="df6bde3b4f1194874df74ce8620448010ffad5b9f135d449273bdad71a069f43" exitCode=0 Dec 27 07:57:56 crc kubenswrapper[4934]: I1227 07:57:56.797995 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wmzfj" event={"ID":"412e15a1-b444-4187-a639-a171ff9a2732","Type":"ContainerDied","Data":"df6bde3b4f1194874df74ce8620448010ffad5b9f135d449273bdad71a069f43"} Dec 27 07:57:57 crc kubenswrapper[4934]: I1227 07:57:57.806546 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wmzfj" event={"ID":"412e15a1-b444-4187-a639-a171ff9a2732","Type":"ContainerStarted","Data":"d0350380713659de84d99fe1f6df2450eda2b7c54458474161a6479796481d87"} Dec 27 07:57:57 crc kubenswrapper[4934]: I1227 07:57:57.826387 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wmzfj" podStartSLOduration=2.35858176 podStartE2EDuration="4.826371947s" podCreationTimestamp="2025-12-27 07:57:53 +0000 UTC" firstStartedPulling="2025-12-27 07:57:54.773893132 +0000 UTC m=+935.594333726" lastFinishedPulling="2025-12-27 07:57:57.241683309 +0000 UTC m=+938.062123913" observedRunningTime="2025-12-27 07:57:57.821750361 +0000 UTC m=+938.642190965" watchObservedRunningTime="2025-12-27 07:57:57.826371947 +0000 UTC m=+938.646812541" Dec 27 07:58:03 crc kubenswrapper[4934]: I1227 07:58:03.553686 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wmzfj" Dec 27 07:58:03 crc kubenswrapper[4934]: I1227 07:58:03.554294 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wmzfj" Dec 27 07:58:03 crc kubenswrapper[4934]: I1227 07:58:03.612180 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wmzfj" Dec 27 07:58:03 crc kubenswrapper[4934]: I1227 07:58:03.916640 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wmzfj" Dec 27 07:58:03 crc kubenswrapper[4934]: I1227 07:58:03.963367 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wmzfj"] Dec 27 07:58:05 crc kubenswrapper[4934]: I1227 07:58:05.885222 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wmzfj" podUID="412e15a1-b444-4187-a639-a171ff9a2732" containerName="registry-server" containerID="cri-o://d0350380713659de84d99fe1f6df2450eda2b7c54458474161a6479796481d87" gracePeriod=2 Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.386835 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wmzfj" Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.450153 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/412e15a1-b444-4187-a639-a171ff9a2732-catalog-content\") pod \"412e15a1-b444-4187-a639-a171ff9a2732\" (UID: \"412e15a1-b444-4187-a639-a171ff9a2732\") " Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.450216 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/412e15a1-b444-4187-a639-a171ff9a2732-utilities\") pod \"412e15a1-b444-4187-a639-a171ff9a2732\" (UID: \"412e15a1-b444-4187-a639-a171ff9a2732\") " Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.451478 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9d8h\" (UniqueName: \"kubernetes.io/projected/412e15a1-b444-4187-a639-a171ff9a2732-kube-api-access-j9d8h\") pod \"412e15a1-b444-4187-a639-a171ff9a2732\" (UID: \"412e15a1-b444-4187-a639-a171ff9a2732\") " Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.451582 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/412e15a1-b444-4187-a639-a171ff9a2732-utilities" (OuterVolumeSpecName: "utilities") pod "412e15a1-b444-4187-a639-a171ff9a2732" (UID: "412e15a1-b444-4187-a639-a171ff9a2732"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.451933 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/412e15a1-b444-4187-a639-a171ff9a2732-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.464518 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/412e15a1-b444-4187-a639-a171ff9a2732-kube-api-access-j9d8h" (OuterVolumeSpecName: "kube-api-access-j9d8h") pod "412e15a1-b444-4187-a639-a171ff9a2732" (UID: "412e15a1-b444-4187-a639-a171ff9a2732"). InnerVolumeSpecName "kube-api-access-j9d8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.555977 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9d8h\" (UniqueName: \"kubernetes.io/projected/412e15a1-b444-4187-a639-a171ff9a2732-kube-api-access-j9d8h\") on node \"crc\" DevicePath \"\"" Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.672167 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/412e15a1-b444-4187-a639-a171ff9a2732-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "412e15a1-b444-4187-a639-a171ff9a2732" (UID: "412e15a1-b444-4187-a639-a171ff9a2732"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.765806 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/412e15a1-b444-4187-a639-a171ff9a2732-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.892627 4934 generic.go:334] "Generic (PLEG): container finished" podID="412e15a1-b444-4187-a639-a171ff9a2732" containerID="d0350380713659de84d99fe1f6df2450eda2b7c54458474161a6479796481d87" exitCode=0 Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.892664 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wmzfj" event={"ID":"412e15a1-b444-4187-a639-a171ff9a2732","Type":"ContainerDied","Data":"d0350380713659de84d99fe1f6df2450eda2b7c54458474161a6479796481d87"} Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.892688 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wmzfj" event={"ID":"412e15a1-b444-4187-a639-a171ff9a2732","Type":"ContainerDied","Data":"d4531431cc5816f9713360c9c2f8251f9321e6fdd9720fab35d0b59580c28fdc"} Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.892691 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wmzfj" Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.892703 4934 scope.go:117] "RemoveContainer" containerID="d0350380713659de84d99fe1f6df2450eda2b7c54458474161a6479796481d87" Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.922562 4934 scope.go:117] "RemoveContainer" containerID="df6bde3b4f1194874df74ce8620448010ffad5b9f135d449273bdad71a069f43" Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.960227 4934 scope.go:117] "RemoveContainer" containerID="82c81277286d3969b89b1ad968d65e0858aed21f718abd35f37b3e820efe3c60" Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.960398 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wmzfj"] Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.965766 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wmzfj"] Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.991174 4934 scope.go:117] "RemoveContainer" containerID="d0350380713659de84d99fe1f6df2450eda2b7c54458474161a6479796481d87" Dec 27 07:58:06 crc kubenswrapper[4934]: E1227 07:58:06.991828 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0350380713659de84d99fe1f6df2450eda2b7c54458474161a6479796481d87\": container with ID starting with d0350380713659de84d99fe1f6df2450eda2b7c54458474161a6479796481d87 not found: ID does not exist" containerID="d0350380713659de84d99fe1f6df2450eda2b7c54458474161a6479796481d87" Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.991968 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0350380713659de84d99fe1f6df2450eda2b7c54458474161a6479796481d87"} err="failed to get container status \"d0350380713659de84d99fe1f6df2450eda2b7c54458474161a6479796481d87\": rpc error: code = NotFound desc = could not find container \"d0350380713659de84d99fe1f6df2450eda2b7c54458474161a6479796481d87\": container with ID starting with d0350380713659de84d99fe1f6df2450eda2b7c54458474161a6479796481d87 not found: ID does not exist" Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.992004 4934 scope.go:117] "RemoveContainer" containerID="df6bde3b4f1194874df74ce8620448010ffad5b9f135d449273bdad71a069f43" Dec 27 07:58:06 crc kubenswrapper[4934]: E1227 07:58:06.992440 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df6bde3b4f1194874df74ce8620448010ffad5b9f135d449273bdad71a069f43\": container with ID starting with df6bde3b4f1194874df74ce8620448010ffad5b9f135d449273bdad71a069f43 not found: ID does not exist" containerID="df6bde3b4f1194874df74ce8620448010ffad5b9f135d449273bdad71a069f43" Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.992494 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df6bde3b4f1194874df74ce8620448010ffad5b9f135d449273bdad71a069f43"} err="failed to get container status \"df6bde3b4f1194874df74ce8620448010ffad5b9f135d449273bdad71a069f43\": rpc error: code = NotFound desc = could not find container \"df6bde3b4f1194874df74ce8620448010ffad5b9f135d449273bdad71a069f43\": container with ID starting with df6bde3b4f1194874df74ce8620448010ffad5b9f135d449273bdad71a069f43 not found: ID does not exist" Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.992530 4934 scope.go:117] "RemoveContainer" containerID="82c81277286d3969b89b1ad968d65e0858aed21f718abd35f37b3e820efe3c60" Dec 27 07:58:06 crc kubenswrapper[4934]: E1227 07:58:06.992867 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82c81277286d3969b89b1ad968d65e0858aed21f718abd35f37b3e820efe3c60\": container with ID starting with 82c81277286d3969b89b1ad968d65e0858aed21f718abd35f37b3e820efe3c60 not found: ID does not exist" containerID="82c81277286d3969b89b1ad968d65e0858aed21f718abd35f37b3e820efe3c60" Dec 27 07:58:06 crc kubenswrapper[4934]: I1227 07:58:06.992909 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82c81277286d3969b89b1ad968d65e0858aed21f718abd35f37b3e820efe3c60"} err="failed to get container status \"82c81277286d3969b89b1ad968d65e0858aed21f718abd35f37b3e820efe3c60\": rpc error: code = NotFound desc = could not find container \"82c81277286d3969b89b1ad968d65e0858aed21f718abd35f37b3e820efe3c60\": container with ID starting with 82c81277286d3969b89b1ad968d65e0858aed21f718abd35f37b3e820efe3c60 not found: ID does not exist" Dec 27 07:58:07 crc kubenswrapper[4934]: I1227 07:58:07.477212 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="412e15a1-b444-4187-a639-a171ff9a2732" path="/var/lib/kubelet/pods/412e15a1-b444-4187-a639-a171ff9a2732/volumes" Dec 27 07:58:07 crc kubenswrapper[4934]: I1227 07:58:07.772621 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-5d9c4bb489-5mh5q" podUID="04fdcc85-c402-43ed-b1bc-4a11307e1640" containerName="console" containerID="cri-o://959bc9a219aee379abea83916b003f39b7a449d9588f9c1cbf8edaf25bb99a38" gracePeriod=15 Dec 27 07:58:08 crc kubenswrapper[4934]: I1227 07:58:08.931870 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5d9c4bb489-5mh5q_04fdcc85-c402-43ed-b1bc-4a11307e1640/console/0.log" Dec 27 07:58:08 crc kubenswrapper[4934]: I1227 07:58:08.932524 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5d9c4bb489-5mh5q" event={"ID":"04fdcc85-c402-43ed-b1bc-4a11307e1640","Type":"ContainerDied","Data":"959bc9a219aee379abea83916b003f39b7a449d9588f9c1cbf8edaf25bb99a38"} Dec 27 07:58:08 crc kubenswrapper[4934]: I1227 07:58:08.932580 4934 generic.go:334] "Generic (PLEG): container finished" podID="04fdcc85-c402-43ed-b1bc-4a11307e1640" containerID="959bc9a219aee379abea83916b003f39b7a449d9588f9c1cbf8edaf25bb99a38" exitCode=2 Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.316861 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5d9c4bb489-5mh5q_04fdcc85-c402-43ed-b1bc-4a11307e1640/console/0.log" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.316940 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.415642 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfz5v\" (UniqueName: \"kubernetes.io/projected/04fdcc85-c402-43ed-b1bc-4a11307e1640-kube-api-access-nfz5v\") pod \"04fdcc85-c402-43ed-b1bc-4a11307e1640\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.415742 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/04fdcc85-c402-43ed-b1bc-4a11307e1640-console-serving-cert\") pod \"04fdcc85-c402-43ed-b1bc-4a11307e1640\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.415779 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-trusted-ca-bundle\") pod \"04fdcc85-c402-43ed-b1bc-4a11307e1640\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.415833 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-oauth-serving-cert\") pod \"04fdcc85-c402-43ed-b1bc-4a11307e1640\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.415866 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-console-config\") pod \"04fdcc85-c402-43ed-b1bc-4a11307e1640\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.415900 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/04fdcc85-c402-43ed-b1bc-4a11307e1640-console-oauth-config\") pod \"04fdcc85-c402-43ed-b1bc-4a11307e1640\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.415980 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-service-ca\") pod \"04fdcc85-c402-43ed-b1bc-4a11307e1640\" (UID: \"04fdcc85-c402-43ed-b1bc-4a11307e1640\") " Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.416811 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-service-ca" (OuterVolumeSpecName: "service-ca") pod "04fdcc85-c402-43ed-b1bc-4a11307e1640" (UID: "04fdcc85-c402-43ed-b1bc-4a11307e1640"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.416843 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-console-config" (OuterVolumeSpecName: "console-config") pod "04fdcc85-c402-43ed-b1bc-4a11307e1640" (UID: "04fdcc85-c402-43ed-b1bc-4a11307e1640"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.417086 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "04fdcc85-c402-43ed-b1bc-4a11307e1640" (UID: "04fdcc85-c402-43ed-b1bc-4a11307e1640"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.417227 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "04fdcc85-c402-43ed-b1bc-4a11307e1640" (UID: "04fdcc85-c402-43ed-b1bc-4a11307e1640"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.421556 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04fdcc85-c402-43ed-b1bc-4a11307e1640-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "04fdcc85-c402-43ed-b1bc-4a11307e1640" (UID: "04fdcc85-c402-43ed-b1bc-4a11307e1640"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.421738 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04fdcc85-c402-43ed-b1bc-4a11307e1640-kube-api-access-nfz5v" (OuterVolumeSpecName: "kube-api-access-nfz5v") pod "04fdcc85-c402-43ed-b1bc-4a11307e1640" (UID: "04fdcc85-c402-43ed-b1bc-4a11307e1640"). InnerVolumeSpecName "kube-api-access-nfz5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.423141 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04fdcc85-c402-43ed-b1bc-4a11307e1640-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "04fdcc85-c402-43ed-b1bc-4a11307e1640" (UID: "04fdcc85-c402-43ed-b1bc-4a11307e1640"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.518370 4934 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/04fdcc85-c402-43ed-b1bc-4a11307e1640-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.518732 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.518755 4934 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.518776 4934 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-console-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.518796 4934 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/04fdcc85-c402-43ed-b1bc-4a11307e1640-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.518814 4934 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/04fdcc85-c402-43ed-b1bc-4a11307e1640-service-ca\") on node \"crc\" DevicePath \"\"" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.518836 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfz5v\" (UniqueName: \"kubernetes.io/projected/04fdcc85-c402-43ed-b1bc-4a11307e1640-kube-api-access-nfz5v\") on node \"crc\" DevicePath \"\"" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.730794 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zr746"] Dec 27 07:58:09 crc kubenswrapper[4934]: E1227 07:58:09.731063 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412e15a1-b444-4187-a639-a171ff9a2732" containerName="registry-server" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.731076 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="412e15a1-b444-4187-a639-a171ff9a2732" containerName="registry-server" Dec 27 07:58:09 crc kubenswrapper[4934]: E1227 07:58:09.731106 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04fdcc85-c402-43ed-b1bc-4a11307e1640" containerName="console" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.731112 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="04fdcc85-c402-43ed-b1bc-4a11307e1640" containerName="console" Dec 27 07:58:09 crc kubenswrapper[4934]: E1227 07:58:09.731126 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412e15a1-b444-4187-a639-a171ff9a2732" containerName="extract-utilities" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.731132 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="412e15a1-b444-4187-a639-a171ff9a2732" containerName="extract-utilities" Dec 27 07:58:09 crc kubenswrapper[4934]: E1227 07:58:09.731140 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="412e15a1-b444-4187-a639-a171ff9a2732" containerName="extract-content" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.731146 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="412e15a1-b444-4187-a639-a171ff9a2732" containerName="extract-content" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.731286 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="04fdcc85-c402-43ed-b1bc-4a11307e1640" containerName="console" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.731301 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="412e15a1-b444-4187-a639-a171ff9a2732" containerName="registry-server" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.733498 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zr746" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.749198 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zr746"] Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.824902 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73efc7d2-bd6d-40e8-9a21-243864fb817b-utilities\") pod \"redhat-marketplace-zr746\" (UID: \"73efc7d2-bd6d-40e8-9a21-243864fb817b\") " pod="openshift-marketplace/redhat-marketplace-zr746" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.825185 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73efc7d2-bd6d-40e8-9a21-243864fb817b-catalog-content\") pod \"redhat-marketplace-zr746\" (UID: \"73efc7d2-bd6d-40e8-9a21-243864fb817b\") " pod="openshift-marketplace/redhat-marketplace-zr746" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.825308 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4p8f\" (UniqueName: \"kubernetes.io/projected/73efc7d2-bd6d-40e8-9a21-243864fb817b-kube-api-access-z4p8f\") pod \"redhat-marketplace-zr746\" (UID: \"73efc7d2-bd6d-40e8-9a21-243864fb817b\") " pod="openshift-marketplace/redhat-marketplace-zr746" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.926882 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73efc7d2-bd6d-40e8-9a21-243864fb817b-catalog-content\") pod \"redhat-marketplace-zr746\" (UID: \"73efc7d2-bd6d-40e8-9a21-243864fb817b\") " pod="openshift-marketplace/redhat-marketplace-zr746" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.926974 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4p8f\" (UniqueName: \"kubernetes.io/projected/73efc7d2-bd6d-40e8-9a21-243864fb817b-kube-api-access-z4p8f\") pod \"redhat-marketplace-zr746\" (UID: \"73efc7d2-bd6d-40e8-9a21-243864fb817b\") " pod="openshift-marketplace/redhat-marketplace-zr746" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.927011 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73efc7d2-bd6d-40e8-9a21-243864fb817b-utilities\") pod \"redhat-marketplace-zr746\" (UID: \"73efc7d2-bd6d-40e8-9a21-243864fb817b\") " pod="openshift-marketplace/redhat-marketplace-zr746" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.927524 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73efc7d2-bd6d-40e8-9a21-243864fb817b-catalog-content\") pod \"redhat-marketplace-zr746\" (UID: \"73efc7d2-bd6d-40e8-9a21-243864fb817b\") " pod="openshift-marketplace/redhat-marketplace-zr746" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.927537 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73efc7d2-bd6d-40e8-9a21-243864fb817b-utilities\") pod \"redhat-marketplace-zr746\" (UID: \"73efc7d2-bd6d-40e8-9a21-243864fb817b\") " pod="openshift-marketplace/redhat-marketplace-zr746" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.940934 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-5d9c4bb489-5mh5q_04fdcc85-c402-43ed-b1bc-4a11307e1640/console/0.log" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.941001 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5d9c4bb489-5mh5q" event={"ID":"04fdcc85-c402-43ed-b1bc-4a11307e1640","Type":"ContainerDied","Data":"116de63b860bbe6f2887b3cf53331715a50d30e24f9ad06ec03115ba449a5d21"} Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.941050 4934 scope.go:117] "RemoveContainer" containerID="959bc9a219aee379abea83916b003f39b7a449d9588f9c1cbf8edaf25bb99a38" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.941076 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5d9c4bb489-5mh5q" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.971158 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4p8f\" (UniqueName: \"kubernetes.io/projected/73efc7d2-bd6d-40e8-9a21-243864fb817b-kube-api-access-z4p8f\") pod \"redhat-marketplace-zr746\" (UID: \"73efc7d2-bd6d-40e8-9a21-243864fb817b\") " pod="openshift-marketplace/redhat-marketplace-zr746" Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.987471 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-5d9c4bb489-5mh5q"] Dec 27 07:58:09 crc kubenswrapper[4934]: I1227 07:58:09.993061 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-5d9c4bb489-5mh5q"] Dec 27 07:58:10 crc kubenswrapper[4934]: I1227 07:58:10.052864 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zr746" Dec 27 07:58:10 crc kubenswrapper[4934]: I1227 07:58:10.383480 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zr746"] Dec 27 07:58:10 crc kubenswrapper[4934]: I1227 07:58:10.897311 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x"] Dec 27 07:58:10 crc kubenswrapper[4934]: I1227 07:58:10.900002 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x" Dec 27 07:58:10 crc kubenswrapper[4934]: I1227 07:58:10.902416 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 27 07:58:10 crc kubenswrapper[4934]: I1227 07:58:10.904770 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x"] Dec 27 07:58:10 crc kubenswrapper[4934]: I1227 07:58:10.951533 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75746381-e0b6-4ad6-8238-ea5f55cb7f54-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x\" (UID: \"75746381-e0b6-4ad6-8238-ea5f55cb7f54\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x" Dec 27 07:58:10 crc kubenswrapper[4934]: I1227 07:58:10.951610 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk2m6\" (UniqueName: \"kubernetes.io/projected/75746381-e0b6-4ad6-8238-ea5f55cb7f54-kube-api-access-nk2m6\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x\" (UID: \"75746381-e0b6-4ad6-8238-ea5f55cb7f54\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x" Dec 27 07:58:10 crc kubenswrapper[4934]: I1227 07:58:10.951802 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75746381-e0b6-4ad6-8238-ea5f55cb7f54-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x\" (UID: \"75746381-e0b6-4ad6-8238-ea5f55cb7f54\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x" Dec 27 07:58:10 crc kubenswrapper[4934]: I1227 07:58:10.956759 4934 generic.go:334] "Generic (PLEG): container finished" podID="73efc7d2-bd6d-40e8-9a21-243864fb817b" containerID="890fdc66ac2ed270e5731554822f8fe712d96295c35ba48a39d34a35f23b6866" exitCode=0 Dec 27 07:58:10 crc kubenswrapper[4934]: I1227 07:58:10.956834 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zr746" event={"ID":"73efc7d2-bd6d-40e8-9a21-243864fb817b","Type":"ContainerDied","Data":"890fdc66ac2ed270e5731554822f8fe712d96295c35ba48a39d34a35f23b6866"} Dec 27 07:58:10 crc kubenswrapper[4934]: I1227 07:58:10.956866 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zr746" event={"ID":"73efc7d2-bd6d-40e8-9a21-243864fb817b","Type":"ContainerStarted","Data":"95525eb7c91172d78dd428cb535394b1b74fc778ee5a6184ad8acb8540290ea4"} Dec 27 07:58:11 crc kubenswrapper[4934]: I1227 07:58:11.053900 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75746381-e0b6-4ad6-8238-ea5f55cb7f54-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x\" (UID: \"75746381-e0b6-4ad6-8238-ea5f55cb7f54\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x" Dec 27 07:58:11 crc kubenswrapper[4934]: I1227 07:58:11.054015 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk2m6\" (UniqueName: \"kubernetes.io/projected/75746381-e0b6-4ad6-8238-ea5f55cb7f54-kube-api-access-nk2m6\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x\" (UID: \"75746381-e0b6-4ad6-8238-ea5f55cb7f54\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x" Dec 27 07:58:11 crc kubenswrapper[4934]: I1227 07:58:11.054233 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75746381-e0b6-4ad6-8238-ea5f55cb7f54-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x\" (UID: \"75746381-e0b6-4ad6-8238-ea5f55cb7f54\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x" Dec 27 07:58:11 crc kubenswrapper[4934]: I1227 07:58:11.054461 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75746381-e0b6-4ad6-8238-ea5f55cb7f54-util\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x\" (UID: \"75746381-e0b6-4ad6-8238-ea5f55cb7f54\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x" Dec 27 07:58:11 crc kubenswrapper[4934]: I1227 07:58:11.054745 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75746381-e0b6-4ad6-8238-ea5f55cb7f54-bundle\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x\" (UID: \"75746381-e0b6-4ad6-8238-ea5f55cb7f54\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x" Dec 27 07:58:11 crc kubenswrapper[4934]: I1227 07:58:11.075370 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk2m6\" (UniqueName: \"kubernetes.io/projected/75746381-e0b6-4ad6-8238-ea5f55cb7f54-kube-api-access-nk2m6\") pod \"5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x\" (UID: \"75746381-e0b6-4ad6-8238-ea5f55cb7f54\") " pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x" Dec 27 07:58:11 crc kubenswrapper[4934]: I1227 07:58:11.213505 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x" Dec 27 07:58:11 crc kubenswrapper[4934]: I1227 07:58:11.479388 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04fdcc85-c402-43ed-b1bc-4a11307e1640" path="/var/lib/kubelet/pods/04fdcc85-c402-43ed-b1bc-4a11307e1640/volumes" Dec 27 07:58:11 crc kubenswrapper[4934]: I1227 07:58:11.676168 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x"] Dec 27 07:58:11 crc kubenswrapper[4934]: W1227 07:58:11.680698 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75746381_e0b6_4ad6_8238_ea5f55cb7f54.slice/crio-1a007338765d3924b31f0421b89fc0df76c70383c069902580be4a59e26f1841 WatchSource:0}: Error finding container 1a007338765d3924b31f0421b89fc0df76c70383c069902580be4a59e26f1841: Status 404 returned error can't find the container with id 1a007338765d3924b31f0421b89fc0df76c70383c069902580be4a59e26f1841 Dec 27 07:58:11 crc kubenswrapper[4934]: I1227 07:58:11.967110 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zr746" event={"ID":"73efc7d2-bd6d-40e8-9a21-243864fb817b","Type":"ContainerStarted","Data":"5e7a9a474decbb983af70eceed212120cfae6c0003a803cd60af66b830ea7b09"} Dec 27 07:58:11 crc kubenswrapper[4934]: I1227 07:58:11.968767 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x" event={"ID":"75746381-e0b6-4ad6-8238-ea5f55cb7f54","Type":"ContainerStarted","Data":"36c2e7d709060b690885178039beba5c19ddffbb4be403f4c8e8684b24b54b78"} Dec 27 07:58:11 crc kubenswrapper[4934]: I1227 07:58:11.968805 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x" event={"ID":"75746381-e0b6-4ad6-8238-ea5f55cb7f54","Type":"ContainerStarted","Data":"1a007338765d3924b31f0421b89fc0df76c70383c069902580be4a59e26f1841"} Dec 27 07:58:12 crc kubenswrapper[4934]: I1227 07:58:12.980353 4934 generic.go:334] "Generic (PLEG): container finished" podID="73efc7d2-bd6d-40e8-9a21-243864fb817b" containerID="5e7a9a474decbb983af70eceed212120cfae6c0003a803cd60af66b830ea7b09" exitCode=0 Dec 27 07:58:12 crc kubenswrapper[4934]: I1227 07:58:12.980474 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zr746" event={"ID":"73efc7d2-bd6d-40e8-9a21-243864fb817b","Type":"ContainerDied","Data":"5e7a9a474decbb983af70eceed212120cfae6c0003a803cd60af66b830ea7b09"} Dec 27 07:58:12 crc kubenswrapper[4934]: I1227 07:58:12.982532 4934 generic.go:334] "Generic (PLEG): container finished" podID="75746381-e0b6-4ad6-8238-ea5f55cb7f54" containerID="36c2e7d709060b690885178039beba5c19ddffbb4be403f4c8e8684b24b54b78" exitCode=0 Dec 27 07:58:12 crc kubenswrapper[4934]: I1227 07:58:12.982590 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x" event={"ID":"75746381-e0b6-4ad6-8238-ea5f55cb7f54","Type":"ContainerDied","Data":"36c2e7d709060b690885178039beba5c19ddffbb4be403f4c8e8684b24b54b78"} Dec 27 07:58:13 crc kubenswrapper[4934]: I1227 07:58:13.994677 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zr746" event={"ID":"73efc7d2-bd6d-40e8-9a21-243864fb817b","Type":"ContainerStarted","Data":"be3961132054857977a2394d293d9bde11fa2335f66b247157dc1dfb2661721d"} Dec 27 07:58:14 crc kubenswrapper[4934]: I1227 07:58:14.022327 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zr746" podStartSLOduration=2.57236251 podStartE2EDuration="5.022309515s" podCreationTimestamp="2025-12-27 07:58:09 +0000 UTC" firstStartedPulling="2025-12-27 07:58:10.960189297 +0000 UTC m=+951.780629901" lastFinishedPulling="2025-12-27 07:58:13.410136312 +0000 UTC m=+954.230576906" observedRunningTime="2025-12-27 07:58:14.017022921 +0000 UTC m=+954.837463525" watchObservedRunningTime="2025-12-27 07:58:14.022309515 +0000 UTC m=+954.842750109" Dec 27 07:58:15 crc kubenswrapper[4934]: I1227 07:58:15.002753 4934 generic.go:334] "Generic (PLEG): container finished" podID="75746381-e0b6-4ad6-8238-ea5f55cb7f54" containerID="b60cb2e9c7123a787eeb32cb183fefeabafca52c899da9726f9264471dffa718" exitCode=0 Dec 27 07:58:15 crc kubenswrapper[4934]: I1227 07:58:15.002839 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x" event={"ID":"75746381-e0b6-4ad6-8238-ea5f55cb7f54","Type":"ContainerDied","Data":"b60cb2e9c7123a787eeb32cb183fefeabafca52c899da9726f9264471dffa718"} Dec 27 07:58:15 crc kubenswrapper[4934]: I1227 07:58:15.330137 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 07:58:15 crc kubenswrapper[4934]: I1227 07:58:15.330219 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 07:58:16 crc kubenswrapper[4934]: I1227 07:58:16.027056 4934 generic.go:334] "Generic (PLEG): container finished" podID="75746381-e0b6-4ad6-8238-ea5f55cb7f54" containerID="ff162ceb8ccc3b3d103ae7e16168cc3199255c81f7ddba6fc3916feb29585015" exitCode=0 Dec 27 07:58:16 crc kubenswrapper[4934]: I1227 07:58:16.027182 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x" event={"ID":"75746381-e0b6-4ad6-8238-ea5f55cb7f54","Type":"ContainerDied","Data":"ff162ceb8ccc3b3d103ae7e16168cc3199255c81f7ddba6fc3916feb29585015"} Dec 27 07:58:17 crc kubenswrapper[4934]: I1227 07:58:17.410131 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x" Dec 27 07:58:17 crc kubenswrapper[4934]: I1227 07:58:17.472876 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75746381-e0b6-4ad6-8238-ea5f55cb7f54-util\") pod \"75746381-e0b6-4ad6-8238-ea5f55cb7f54\" (UID: \"75746381-e0b6-4ad6-8238-ea5f55cb7f54\") " Dec 27 07:58:17 crc kubenswrapper[4934]: I1227 07:58:17.473041 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75746381-e0b6-4ad6-8238-ea5f55cb7f54-bundle\") pod \"75746381-e0b6-4ad6-8238-ea5f55cb7f54\" (UID: \"75746381-e0b6-4ad6-8238-ea5f55cb7f54\") " Dec 27 07:58:17 crc kubenswrapper[4934]: I1227 07:58:17.473138 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nk2m6\" (UniqueName: \"kubernetes.io/projected/75746381-e0b6-4ad6-8238-ea5f55cb7f54-kube-api-access-nk2m6\") pod \"75746381-e0b6-4ad6-8238-ea5f55cb7f54\" (UID: \"75746381-e0b6-4ad6-8238-ea5f55cb7f54\") " Dec 27 07:58:17 crc kubenswrapper[4934]: I1227 07:58:17.474565 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75746381-e0b6-4ad6-8238-ea5f55cb7f54-bundle" (OuterVolumeSpecName: "bundle") pod "75746381-e0b6-4ad6-8238-ea5f55cb7f54" (UID: "75746381-e0b6-4ad6-8238-ea5f55cb7f54"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:58:17 crc kubenswrapper[4934]: I1227 07:58:17.490317 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75746381-e0b6-4ad6-8238-ea5f55cb7f54-kube-api-access-nk2m6" (OuterVolumeSpecName: "kube-api-access-nk2m6") pod "75746381-e0b6-4ad6-8238-ea5f55cb7f54" (UID: "75746381-e0b6-4ad6-8238-ea5f55cb7f54"). InnerVolumeSpecName "kube-api-access-nk2m6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:58:17 crc kubenswrapper[4934]: I1227 07:58:17.574967 4934 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75746381-e0b6-4ad6-8238-ea5f55cb7f54-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 07:58:17 crc kubenswrapper[4934]: I1227 07:58:17.575024 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nk2m6\" (UniqueName: \"kubernetes.io/projected/75746381-e0b6-4ad6-8238-ea5f55cb7f54-kube-api-access-nk2m6\") on node \"crc\" DevicePath \"\"" Dec 27 07:58:17 crc kubenswrapper[4934]: I1227 07:58:17.851058 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75746381-e0b6-4ad6-8238-ea5f55cb7f54-util" (OuterVolumeSpecName: "util") pod "75746381-e0b6-4ad6-8238-ea5f55cb7f54" (UID: "75746381-e0b6-4ad6-8238-ea5f55cb7f54"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:58:17 crc kubenswrapper[4934]: I1227 07:58:17.879814 4934 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75746381-e0b6-4ad6-8238-ea5f55cb7f54-util\") on node \"crc\" DevicePath \"\"" Dec 27 07:58:18 crc kubenswrapper[4934]: I1227 07:58:18.044299 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x" event={"ID":"75746381-e0b6-4ad6-8238-ea5f55cb7f54","Type":"ContainerDied","Data":"1a007338765d3924b31f0421b89fc0df76c70383c069902580be4a59e26f1841"} Dec 27 07:58:18 crc kubenswrapper[4934]: I1227 07:58:18.044336 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a007338765d3924b31f0421b89fc0df76c70383c069902580be4a59e26f1841" Dec 27 07:58:18 crc kubenswrapper[4934]: I1227 07:58:18.044362 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x" Dec 27 07:58:20 crc kubenswrapper[4934]: I1227 07:58:20.053827 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zr746" Dec 27 07:58:20 crc kubenswrapper[4934]: I1227 07:58:20.053858 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zr746" Dec 27 07:58:20 crc kubenswrapper[4934]: I1227 07:58:20.108992 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zr746" Dec 27 07:58:21 crc kubenswrapper[4934]: I1227 07:58:21.157769 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zr746" Dec 27 07:58:21 crc kubenswrapper[4934]: I1227 07:58:21.458470 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zr746"] Dec 27 07:58:23 crc kubenswrapper[4934]: I1227 07:58:23.078715 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zr746" podUID="73efc7d2-bd6d-40e8-9a21-243864fb817b" containerName="registry-server" containerID="cri-o://be3961132054857977a2394d293d9bde11fa2335f66b247157dc1dfb2661721d" gracePeriod=2 Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.026676 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zr746" Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.081831 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4p8f\" (UniqueName: \"kubernetes.io/projected/73efc7d2-bd6d-40e8-9a21-243864fb817b-kube-api-access-z4p8f\") pod \"73efc7d2-bd6d-40e8-9a21-243864fb817b\" (UID: \"73efc7d2-bd6d-40e8-9a21-243864fb817b\") " Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.081926 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73efc7d2-bd6d-40e8-9a21-243864fb817b-catalog-content\") pod \"73efc7d2-bd6d-40e8-9a21-243864fb817b\" (UID: \"73efc7d2-bd6d-40e8-9a21-243864fb817b\") " Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.082075 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73efc7d2-bd6d-40e8-9a21-243864fb817b-utilities\") pod \"73efc7d2-bd6d-40e8-9a21-243864fb817b\" (UID: \"73efc7d2-bd6d-40e8-9a21-243864fb817b\") " Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.088832 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73efc7d2-bd6d-40e8-9a21-243864fb817b-utilities" (OuterVolumeSpecName: "utilities") pod "73efc7d2-bd6d-40e8-9a21-243864fb817b" (UID: "73efc7d2-bd6d-40e8-9a21-243864fb817b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.095416 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73efc7d2-bd6d-40e8-9a21-243864fb817b-kube-api-access-z4p8f" (OuterVolumeSpecName: "kube-api-access-z4p8f") pod "73efc7d2-bd6d-40e8-9a21-243864fb817b" (UID: "73efc7d2-bd6d-40e8-9a21-243864fb817b"). InnerVolumeSpecName "kube-api-access-z4p8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.115232 4934 generic.go:334] "Generic (PLEG): container finished" podID="73efc7d2-bd6d-40e8-9a21-243864fb817b" containerID="be3961132054857977a2394d293d9bde11fa2335f66b247157dc1dfb2661721d" exitCode=0 Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.115276 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zr746" event={"ID":"73efc7d2-bd6d-40e8-9a21-243864fb817b","Type":"ContainerDied","Data":"be3961132054857977a2394d293d9bde11fa2335f66b247157dc1dfb2661721d"} Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.115301 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zr746" event={"ID":"73efc7d2-bd6d-40e8-9a21-243864fb817b","Type":"ContainerDied","Data":"95525eb7c91172d78dd428cb535394b1b74fc778ee5a6184ad8acb8540290ea4"} Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.115317 4934 scope.go:117] "RemoveContainer" containerID="be3961132054857977a2394d293d9bde11fa2335f66b247157dc1dfb2661721d" Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.115433 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zr746" Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.138266 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73efc7d2-bd6d-40e8-9a21-243864fb817b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73efc7d2-bd6d-40e8-9a21-243864fb817b" (UID: "73efc7d2-bd6d-40e8-9a21-243864fb817b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.142727 4934 scope.go:117] "RemoveContainer" containerID="5e7a9a474decbb983af70eceed212120cfae6c0003a803cd60af66b830ea7b09" Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.173790 4934 scope.go:117] "RemoveContainer" containerID="890fdc66ac2ed270e5731554822f8fe712d96295c35ba48a39d34a35f23b6866" Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.184884 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4p8f\" (UniqueName: \"kubernetes.io/projected/73efc7d2-bd6d-40e8-9a21-243864fb817b-kube-api-access-z4p8f\") on node \"crc\" DevicePath \"\"" Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.184928 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73efc7d2-bd6d-40e8-9a21-243864fb817b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.184938 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73efc7d2-bd6d-40e8-9a21-243864fb817b-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.187372 4934 scope.go:117] "RemoveContainer" containerID="be3961132054857977a2394d293d9bde11fa2335f66b247157dc1dfb2661721d" Dec 27 07:58:24 crc kubenswrapper[4934]: E1227 07:58:24.191490 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be3961132054857977a2394d293d9bde11fa2335f66b247157dc1dfb2661721d\": container with ID starting with be3961132054857977a2394d293d9bde11fa2335f66b247157dc1dfb2661721d not found: ID does not exist" containerID="be3961132054857977a2394d293d9bde11fa2335f66b247157dc1dfb2661721d" Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.191534 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be3961132054857977a2394d293d9bde11fa2335f66b247157dc1dfb2661721d"} err="failed to get container status \"be3961132054857977a2394d293d9bde11fa2335f66b247157dc1dfb2661721d\": rpc error: code = NotFound desc = could not find container \"be3961132054857977a2394d293d9bde11fa2335f66b247157dc1dfb2661721d\": container with ID starting with be3961132054857977a2394d293d9bde11fa2335f66b247157dc1dfb2661721d not found: ID does not exist" Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.191562 4934 scope.go:117] "RemoveContainer" containerID="5e7a9a474decbb983af70eceed212120cfae6c0003a803cd60af66b830ea7b09" Dec 27 07:58:24 crc kubenswrapper[4934]: E1227 07:58:24.191916 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e7a9a474decbb983af70eceed212120cfae6c0003a803cd60af66b830ea7b09\": container with ID starting with 5e7a9a474decbb983af70eceed212120cfae6c0003a803cd60af66b830ea7b09 not found: ID does not exist" containerID="5e7a9a474decbb983af70eceed212120cfae6c0003a803cd60af66b830ea7b09" Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.191956 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e7a9a474decbb983af70eceed212120cfae6c0003a803cd60af66b830ea7b09"} err="failed to get container status \"5e7a9a474decbb983af70eceed212120cfae6c0003a803cd60af66b830ea7b09\": rpc error: code = NotFound desc = could not find container \"5e7a9a474decbb983af70eceed212120cfae6c0003a803cd60af66b830ea7b09\": container with ID starting with 5e7a9a474decbb983af70eceed212120cfae6c0003a803cd60af66b830ea7b09 not found: ID does not exist" Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.191983 4934 scope.go:117] "RemoveContainer" containerID="890fdc66ac2ed270e5731554822f8fe712d96295c35ba48a39d34a35f23b6866" Dec 27 07:58:24 crc kubenswrapper[4934]: E1227 07:58:24.192385 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"890fdc66ac2ed270e5731554822f8fe712d96295c35ba48a39d34a35f23b6866\": container with ID starting with 890fdc66ac2ed270e5731554822f8fe712d96295c35ba48a39d34a35f23b6866 not found: ID does not exist" containerID="890fdc66ac2ed270e5731554822f8fe712d96295c35ba48a39d34a35f23b6866" Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.192420 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"890fdc66ac2ed270e5731554822f8fe712d96295c35ba48a39d34a35f23b6866"} err="failed to get container status \"890fdc66ac2ed270e5731554822f8fe712d96295c35ba48a39d34a35f23b6866\": rpc error: code = NotFound desc = could not find container \"890fdc66ac2ed270e5731554822f8fe712d96295c35ba48a39d34a35f23b6866\": container with ID starting with 890fdc66ac2ed270e5731554822f8fe712d96295c35ba48a39d34a35f23b6866 not found: ID does not exist" Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.454348 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zr746"] Dec 27 07:58:24 crc kubenswrapper[4934]: I1227 07:58:24.462275 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zr746"] Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.476003 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73efc7d2-bd6d-40e8-9a21-243864fb817b" path="/var/lib/kubelet/pods/73efc7d2-bd6d-40e8-9a21-243864fb817b/volumes" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.702515 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z"] Dec 27 07:58:25 crc kubenswrapper[4934]: E1227 07:58:25.702791 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75746381-e0b6-4ad6-8238-ea5f55cb7f54" containerName="pull" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.702806 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="75746381-e0b6-4ad6-8238-ea5f55cb7f54" containerName="pull" Dec 27 07:58:25 crc kubenswrapper[4934]: E1227 07:58:25.702838 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73efc7d2-bd6d-40e8-9a21-243864fb817b" containerName="extract-utilities" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.702845 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="73efc7d2-bd6d-40e8-9a21-243864fb817b" containerName="extract-utilities" Dec 27 07:58:25 crc kubenswrapper[4934]: E1227 07:58:25.702851 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75746381-e0b6-4ad6-8238-ea5f55cb7f54" containerName="extract" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.702857 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="75746381-e0b6-4ad6-8238-ea5f55cb7f54" containerName="extract" Dec 27 07:58:25 crc kubenswrapper[4934]: E1227 07:58:25.702880 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75746381-e0b6-4ad6-8238-ea5f55cb7f54" containerName="util" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.702886 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="75746381-e0b6-4ad6-8238-ea5f55cb7f54" containerName="util" Dec 27 07:58:25 crc kubenswrapper[4934]: E1227 07:58:25.702912 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73efc7d2-bd6d-40e8-9a21-243864fb817b" containerName="registry-server" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.702920 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="73efc7d2-bd6d-40e8-9a21-243864fb817b" containerName="registry-server" Dec 27 07:58:25 crc kubenswrapper[4934]: E1227 07:58:25.702936 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73efc7d2-bd6d-40e8-9a21-243864fb817b" containerName="extract-content" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.702941 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="73efc7d2-bd6d-40e8-9a21-243864fb817b" containerName="extract-content" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.703067 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="75746381-e0b6-4ad6-8238-ea5f55cb7f54" containerName="extract" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.703092 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="73efc7d2-bd6d-40e8-9a21-243864fb817b" containerName="registry-server" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.703579 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.705767 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.705862 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-p74h8" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.708465 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.708691 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.708853 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.720356 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z"] Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.809696 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107-webhook-cert\") pod \"metallb-operator-controller-manager-5b75b656f5-sqn2z\" (UID: \"fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107\") " pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.809757 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107-apiservice-cert\") pod \"metallb-operator-controller-manager-5b75b656f5-sqn2z\" (UID: \"fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107\") " pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.809877 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrv2h\" (UniqueName: \"kubernetes.io/projected/fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107-kube-api-access-rrv2h\") pod \"metallb-operator-controller-manager-5b75b656f5-sqn2z\" (UID: \"fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107\") " pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.911880 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrv2h\" (UniqueName: \"kubernetes.io/projected/fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107-kube-api-access-rrv2h\") pod \"metallb-operator-controller-manager-5b75b656f5-sqn2z\" (UID: \"fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107\") " pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.911975 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107-webhook-cert\") pod \"metallb-operator-controller-manager-5b75b656f5-sqn2z\" (UID: \"fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107\") " pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.911998 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107-apiservice-cert\") pod \"metallb-operator-controller-manager-5b75b656f5-sqn2z\" (UID: \"fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107\") " pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.916766 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107-webhook-cert\") pod \"metallb-operator-controller-manager-5b75b656f5-sqn2z\" (UID: \"fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107\") " pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.922572 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107-apiservice-cert\") pod \"metallb-operator-controller-manager-5b75b656f5-sqn2z\" (UID: \"fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107\") " pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" Dec 27 07:58:25 crc kubenswrapper[4934]: I1227 07:58:25.927149 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrv2h\" (UniqueName: \"kubernetes.io/projected/fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107-kube-api-access-rrv2h\") pod \"metallb-operator-controller-manager-5b75b656f5-sqn2z\" (UID: \"fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107\") " pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" Dec 27 07:58:26 crc kubenswrapper[4934]: I1227 07:58:26.020123 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" Dec 27 07:58:26 crc kubenswrapper[4934]: I1227 07:58:26.020744 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8"] Dec 27 07:58:26 crc kubenswrapper[4934]: I1227 07:58:26.021760 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" Dec 27 07:58:26 crc kubenswrapper[4934]: I1227 07:58:26.024068 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-5gtz6" Dec 27 07:58:26 crc kubenswrapper[4934]: I1227 07:58:26.024402 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 27 07:58:26 crc kubenswrapper[4934]: I1227 07:58:26.031539 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8"] Dec 27 07:58:26 crc kubenswrapper[4934]: I1227 07:58:26.034434 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 27 07:58:26 crc kubenswrapper[4934]: I1227 07:58:26.115743 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb-apiservice-cert\") pod \"metallb-operator-webhook-server-8b4577cdd-qcwh8\" (UID: \"df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb\") " pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" Dec 27 07:58:26 crc kubenswrapper[4934]: I1227 07:58:26.116132 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qs7lx\" (UniqueName: \"kubernetes.io/projected/df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb-kube-api-access-qs7lx\") pod \"metallb-operator-webhook-server-8b4577cdd-qcwh8\" (UID: \"df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb\") " pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" Dec 27 07:58:26 crc kubenswrapper[4934]: I1227 07:58:26.116173 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb-webhook-cert\") pod \"metallb-operator-webhook-server-8b4577cdd-qcwh8\" (UID: \"df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb\") " pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" Dec 27 07:58:26 crc kubenswrapper[4934]: I1227 07:58:26.219176 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb-apiservice-cert\") pod \"metallb-operator-webhook-server-8b4577cdd-qcwh8\" (UID: \"df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb\") " pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" Dec 27 07:58:26 crc kubenswrapper[4934]: I1227 07:58:26.219240 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qs7lx\" (UniqueName: \"kubernetes.io/projected/df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb-kube-api-access-qs7lx\") pod \"metallb-operator-webhook-server-8b4577cdd-qcwh8\" (UID: \"df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb\") " pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" Dec 27 07:58:26 crc kubenswrapper[4934]: I1227 07:58:26.219265 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb-webhook-cert\") pod \"metallb-operator-webhook-server-8b4577cdd-qcwh8\" (UID: \"df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb\") " pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" Dec 27 07:58:26 crc kubenswrapper[4934]: I1227 07:58:26.223852 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb-apiservice-cert\") pod \"metallb-operator-webhook-server-8b4577cdd-qcwh8\" (UID: \"df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb\") " pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" Dec 27 07:58:26 crc kubenswrapper[4934]: I1227 07:58:26.248892 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qs7lx\" (UniqueName: \"kubernetes.io/projected/df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb-kube-api-access-qs7lx\") pod \"metallb-operator-webhook-server-8b4577cdd-qcwh8\" (UID: \"df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb\") " pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" Dec 27 07:58:26 crc kubenswrapper[4934]: I1227 07:58:26.249701 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb-webhook-cert\") pod \"metallb-operator-webhook-server-8b4577cdd-qcwh8\" (UID: \"df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb\") " pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" Dec 27 07:58:26 crc kubenswrapper[4934]: I1227 07:58:26.343569 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" Dec 27 07:58:26 crc kubenswrapper[4934]: I1227 07:58:26.527747 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z"] Dec 27 07:58:26 crc kubenswrapper[4934]: I1227 07:58:26.760654 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8"] Dec 27 07:58:26 crc kubenswrapper[4934]: W1227 07:58:26.763380 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf8d44b2_c06a_4d4f_8ff8_6bde8f8b4acb.slice/crio-00fc737ffac23ddb09dd62ceefe1fc41ad94a917f2be75fb6a79617679a39d68 WatchSource:0}: Error finding container 00fc737ffac23ddb09dd62ceefe1fc41ad94a917f2be75fb6a79617679a39d68: Status 404 returned error can't find the container with id 00fc737ffac23ddb09dd62ceefe1fc41ad94a917f2be75fb6a79617679a39d68 Dec 27 07:58:27 crc kubenswrapper[4934]: I1227 07:58:27.135669 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" event={"ID":"df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb","Type":"ContainerStarted","Data":"00fc737ffac23ddb09dd62ceefe1fc41ad94a917f2be75fb6a79617679a39d68"} Dec 27 07:58:27 crc kubenswrapper[4934]: I1227 07:58:27.136926 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" event={"ID":"fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107","Type":"ContainerStarted","Data":"c6d23516885711b0602719bf522ce82b57b906ce9da2244291c0d4771aad1271"} Dec 27 07:58:31 crc kubenswrapper[4934]: I1227 07:58:31.177207 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" event={"ID":"fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107","Type":"ContainerStarted","Data":"c5033b01f62f941325bff8f8370a741678487ec1c9e2743d1eb80a50d5c7c8e8"} Dec 27 07:58:31 crc kubenswrapper[4934]: I1227 07:58:31.177722 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" Dec 27 07:58:31 crc kubenswrapper[4934]: I1227 07:58:31.229435 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" podStartSLOduration=2.769720304 podStartE2EDuration="6.229405821s" podCreationTimestamp="2025-12-27 07:58:25 +0000 UTC" firstStartedPulling="2025-12-27 07:58:26.542803453 +0000 UTC m=+967.363244047" lastFinishedPulling="2025-12-27 07:58:30.00248896 +0000 UTC m=+970.822929564" observedRunningTime="2025-12-27 07:58:31.221287276 +0000 UTC m=+972.041727880" watchObservedRunningTime="2025-12-27 07:58:31.229405821 +0000 UTC m=+972.049846415" Dec 27 07:58:34 crc kubenswrapper[4934]: I1227 07:58:34.201726 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" event={"ID":"df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb","Type":"ContainerStarted","Data":"0544eea826fad2f82fea05a46a3bf73c02516d04af2f770e270401a3e238142f"} Dec 27 07:58:34 crc kubenswrapper[4934]: I1227 07:58:34.202037 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" Dec 27 07:58:45 crc kubenswrapper[4934]: I1227 07:58:45.329812 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 07:58:45 crc kubenswrapper[4934]: I1227 07:58:45.330413 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 07:58:45 crc kubenswrapper[4934]: I1227 07:58:45.330458 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 07:58:45 crc kubenswrapper[4934]: I1227 07:58:45.331059 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4ea115bb4d50c7db66f5a9c8846c893a8cb9f55d01ba76e6418fe5095edfd343"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 07:58:45 crc kubenswrapper[4934]: I1227 07:58:45.331128 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://4ea115bb4d50c7db66f5a9c8846c893a8cb9f55d01ba76e6418fe5095edfd343" gracePeriod=600 Dec 27 07:58:46 crc kubenswrapper[4934]: I1227 07:58:46.348494 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" Dec 27 07:58:46 crc kubenswrapper[4934]: I1227 07:58:46.373458 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" podStartSLOduration=13.830532996 podStartE2EDuration="20.373441601s" podCreationTimestamp="2025-12-27 07:58:26 +0000 UTC" firstStartedPulling="2025-12-27 07:58:26.767873981 +0000 UTC m=+967.588314575" lastFinishedPulling="2025-12-27 07:58:33.310782586 +0000 UTC m=+974.131223180" observedRunningTime="2025-12-27 07:58:34.224595948 +0000 UTC m=+975.045036552" watchObservedRunningTime="2025-12-27 07:58:46.373441601 +0000 UTC m=+987.193882195" Dec 27 07:58:46 crc kubenswrapper[4934]: I1227 07:58:46.519323 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="4ea115bb4d50c7db66f5a9c8846c893a8cb9f55d01ba76e6418fe5095edfd343" exitCode=0 Dec 27 07:58:46 crc kubenswrapper[4934]: I1227 07:58:46.519375 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"4ea115bb4d50c7db66f5a9c8846c893a8cb9f55d01ba76e6418fe5095edfd343"} Dec 27 07:58:46 crc kubenswrapper[4934]: I1227 07:58:46.519417 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"b960fbd3418f51dd41277ee06c40ec3eb700b9243e777f61c98b4b5b56bf0c98"} Dec 27 07:58:46 crc kubenswrapper[4934]: I1227 07:58:46.519438 4934 scope.go:117] "RemoveContainer" containerID="56bf8d198d1688bbc71036fbf874230b44fa217f3818e5e1d2117e459b2b5477" Dec 27 07:59:06 crc kubenswrapper[4934]: I1227 07:59:06.024831 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" Dec 27 07:59:06 crc kubenswrapper[4934]: I1227 07:59:06.933726 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-v2lll"] Dec 27 07:59:06 crc kubenswrapper[4934]: I1227 07:59:06.943384 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:06 crc kubenswrapper[4934]: I1227 07:59:06.950377 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 27 07:59:06 crc kubenswrapper[4934]: I1227 07:59:06.950401 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 27 07:59:06 crc kubenswrapper[4934]: I1227 07:59:06.951397 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-256fp" Dec 27 07:59:06 crc kubenswrapper[4934]: I1227 07:59:06.964797 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-5bsq9"] Dec 27 07:59:06 crc kubenswrapper[4934]: I1227 07:59:06.966622 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5bsq9" Dec 27 07:59:06 crc kubenswrapper[4934]: I1227 07:59:06.971610 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 27 07:59:06 crc kubenswrapper[4934]: I1227 07:59:06.987181 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-5bsq9"] Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.000432 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m62k2\" (UniqueName: \"kubernetes.io/projected/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-kube-api-access-m62k2\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.000477 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-metrics-certs\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.000506 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-metrics\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.000549 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-reloader\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.000582 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-frr-conf\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.000619 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-frr-sockets\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.000653 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-frr-startup\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.052758 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-4jwwj"] Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.054076 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-4jwwj" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.058717 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.058926 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.058946 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-dttn4" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.059038 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.075595 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-5bddd4b946-7c5ls"] Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.076833 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5bddd4b946-7c5ls" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.083594 4934 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.087551 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5bddd4b946-7c5ls"] Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.102956 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqx4b\" (UniqueName: \"kubernetes.io/projected/ea01c7df-c5aa-4c86-800b-b6aebafeb8fd-kube-api-access-hqx4b\") pod \"speaker-4jwwj\" (UID: \"ea01c7df-c5aa-4c86-800b-b6aebafeb8fd\") " pod="metallb-system/speaker-4jwwj" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.103017 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ea01c7df-c5aa-4c86-800b-b6aebafeb8fd-metallb-excludel2\") pod \"speaker-4jwwj\" (UID: \"ea01c7df-c5aa-4c86-800b-b6aebafeb8fd\") " pod="metallb-system/speaker-4jwwj" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.103096 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-frr-startup\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.103127 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbtbp\" (UniqueName: \"kubernetes.io/projected/0134c36f-c43b-4830-848a-3a8690957ee1-kube-api-access-bbtbp\") pod \"frr-k8s-webhook-server-7784b6fcf-5bsq9\" (UID: \"0134c36f-c43b-4830-848a-3a8690957ee1\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5bsq9" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.103153 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ea01c7df-c5aa-4c86-800b-b6aebafeb8fd-metrics-certs\") pod \"speaker-4jwwj\" (UID: \"ea01c7df-c5aa-4c86-800b-b6aebafeb8fd\") " pod="metallb-system/speaker-4jwwj" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.103180 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ea01c7df-c5aa-4c86-800b-b6aebafeb8fd-memberlist\") pod \"speaker-4jwwj\" (UID: \"ea01c7df-c5aa-4c86-800b-b6aebafeb8fd\") " pod="metallb-system/speaker-4jwwj" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.103201 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m62k2\" (UniqueName: \"kubernetes.io/projected/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-kube-api-access-m62k2\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.103223 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-metrics-certs\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.103245 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0134c36f-c43b-4830-848a-3a8690957ee1-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-5bsq9\" (UID: \"0134c36f-c43b-4830-848a-3a8690957ee1\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5bsq9" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.103267 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-metrics\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.103286 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-reloader\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.103328 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-frr-conf\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: E1227 07:59:07.103355 4934 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 27 07:59:07 crc kubenswrapper[4934]: E1227 07:59:07.103414 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-metrics-certs podName:c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd nodeName:}" failed. No retries permitted until 2025-12-27 07:59:07.603397719 +0000 UTC m=+1008.423838313 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-metrics-certs") pod "frr-k8s-v2lll" (UID: "c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd") : secret "frr-k8s-certs-secret" not found Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.103369 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-frr-sockets\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.104020 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-frr-conf\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.104069 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-frr-sockets\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.104292 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-reloader\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.104463 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-metrics\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.104628 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-frr-startup\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.127767 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m62k2\" (UniqueName: \"kubernetes.io/projected/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-kube-api-access-m62k2\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.205288 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcgkz\" (UniqueName: \"kubernetes.io/projected/a8825499-558f-4b31-b09b-0dcc92662f65-kube-api-access-bcgkz\") pod \"controller-5bddd4b946-7c5ls\" (UID: \"a8825499-558f-4b31-b09b-0dcc92662f65\") " pod="metallb-system/controller-5bddd4b946-7c5ls" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.205359 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqx4b\" (UniqueName: \"kubernetes.io/projected/ea01c7df-c5aa-4c86-800b-b6aebafeb8fd-kube-api-access-hqx4b\") pod \"speaker-4jwwj\" (UID: \"ea01c7df-c5aa-4c86-800b-b6aebafeb8fd\") " pod="metallb-system/speaker-4jwwj" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.205403 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ea01c7df-c5aa-4c86-800b-b6aebafeb8fd-metallb-excludel2\") pod \"speaker-4jwwj\" (UID: \"ea01c7df-c5aa-4c86-800b-b6aebafeb8fd\") " pod="metallb-system/speaker-4jwwj" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.205455 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbtbp\" (UniqueName: \"kubernetes.io/projected/0134c36f-c43b-4830-848a-3a8690957ee1-kube-api-access-bbtbp\") pod \"frr-k8s-webhook-server-7784b6fcf-5bsq9\" (UID: \"0134c36f-c43b-4830-848a-3a8690957ee1\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5bsq9" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.205481 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ea01c7df-c5aa-4c86-800b-b6aebafeb8fd-metrics-certs\") pod \"speaker-4jwwj\" (UID: \"ea01c7df-c5aa-4c86-800b-b6aebafeb8fd\") " pod="metallb-system/speaker-4jwwj" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.205515 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ea01c7df-c5aa-4c86-800b-b6aebafeb8fd-memberlist\") pod \"speaker-4jwwj\" (UID: \"ea01c7df-c5aa-4c86-800b-b6aebafeb8fd\") " pod="metallb-system/speaker-4jwwj" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.205565 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a8825499-558f-4b31-b09b-0dcc92662f65-metrics-certs\") pod \"controller-5bddd4b946-7c5ls\" (UID: \"a8825499-558f-4b31-b09b-0dcc92662f65\") " pod="metallb-system/controller-5bddd4b946-7c5ls" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.205594 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0134c36f-c43b-4830-848a-3a8690957ee1-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-5bsq9\" (UID: \"0134c36f-c43b-4830-848a-3a8690957ee1\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5bsq9" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.205637 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a8825499-558f-4b31-b09b-0dcc92662f65-cert\") pod \"controller-5bddd4b946-7c5ls\" (UID: \"a8825499-558f-4b31-b09b-0dcc92662f65\") " pod="metallb-system/controller-5bddd4b946-7c5ls" Dec 27 07:59:07 crc kubenswrapper[4934]: E1227 07:59:07.205719 4934 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 27 07:59:07 crc kubenswrapper[4934]: E1227 07:59:07.205776 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ea01c7df-c5aa-4c86-800b-b6aebafeb8fd-metrics-certs podName:ea01c7df-c5aa-4c86-800b-b6aebafeb8fd nodeName:}" failed. No retries permitted until 2025-12-27 07:59:07.705760872 +0000 UTC m=+1008.526201466 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ea01c7df-c5aa-4c86-800b-b6aebafeb8fd-metrics-certs") pod "speaker-4jwwj" (UID: "ea01c7df-c5aa-4c86-800b-b6aebafeb8fd") : secret "speaker-certs-secret" not found Dec 27 07:59:07 crc kubenswrapper[4934]: E1227 07:59:07.205797 4934 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 27 07:59:07 crc kubenswrapper[4934]: E1227 07:59:07.205879 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ea01c7df-c5aa-4c86-800b-b6aebafeb8fd-memberlist podName:ea01c7df-c5aa-4c86-800b-b6aebafeb8fd nodeName:}" failed. No retries permitted until 2025-12-27 07:59:07.705834494 +0000 UTC m=+1008.526275198 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/ea01c7df-c5aa-4c86-800b-b6aebafeb8fd-memberlist") pod "speaker-4jwwj" (UID: "ea01c7df-c5aa-4c86-800b-b6aebafeb8fd") : secret "metallb-memberlist" not found Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.206395 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ea01c7df-c5aa-4c86-800b-b6aebafeb8fd-metallb-excludel2\") pod \"speaker-4jwwj\" (UID: \"ea01c7df-c5aa-4c86-800b-b6aebafeb8fd\") " pod="metallb-system/speaker-4jwwj" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.210504 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0134c36f-c43b-4830-848a-3a8690957ee1-cert\") pod \"frr-k8s-webhook-server-7784b6fcf-5bsq9\" (UID: \"0134c36f-c43b-4830-848a-3a8690957ee1\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5bsq9" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.227616 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqx4b\" (UniqueName: \"kubernetes.io/projected/ea01c7df-c5aa-4c86-800b-b6aebafeb8fd-kube-api-access-hqx4b\") pod \"speaker-4jwwj\" (UID: \"ea01c7df-c5aa-4c86-800b-b6aebafeb8fd\") " pod="metallb-system/speaker-4jwwj" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.234478 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbtbp\" (UniqueName: \"kubernetes.io/projected/0134c36f-c43b-4830-848a-3a8690957ee1-kube-api-access-bbtbp\") pod \"frr-k8s-webhook-server-7784b6fcf-5bsq9\" (UID: \"0134c36f-c43b-4830-848a-3a8690957ee1\") " pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5bsq9" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.292726 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5bsq9" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.307538 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a8825499-558f-4b31-b09b-0dcc92662f65-metrics-certs\") pod \"controller-5bddd4b946-7c5ls\" (UID: \"a8825499-558f-4b31-b09b-0dcc92662f65\") " pod="metallb-system/controller-5bddd4b946-7c5ls" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.307611 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a8825499-558f-4b31-b09b-0dcc92662f65-cert\") pod \"controller-5bddd4b946-7c5ls\" (UID: \"a8825499-558f-4b31-b09b-0dcc92662f65\") " pod="metallb-system/controller-5bddd4b946-7c5ls" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.307671 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcgkz\" (UniqueName: \"kubernetes.io/projected/a8825499-558f-4b31-b09b-0dcc92662f65-kube-api-access-bcgkz\") pod \"controller-5bddd4b946-7c5ls\" (UID: \"a8825499-558f-4b31-b09b-0dcc92662f65\") " pod="metallb-system/controller-5bddd4b946-7c5ls" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.312354 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a8825499-558f-4b31-b09b-0dcc92662f65-cert\") pod \"controller-5bddd4b946-7c5ls\" (UID: \"a8825499-558f-4b31-b09b-0dcc92662f65\") " pod="metallb-system/controller-5bddd4b946-7c5ls" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.313377 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a8825499-558f-4b31-b09b-0dcc92662f65-metrics-certs\") pod \"controller-5bddd4b946-7c5ls\" (UID: \"a8825499-558f-4b31-b09b-0dcc92662f65\") " pod="metallb-system/controller-5bddd4b946-7c5ls" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.325647 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcgkz\" (UniqueName: \"kubernetes.io/projected/a8825499-558f-4b31-b09b-0dcc92662f65-kube-api-access-bcgkz\") pod \"controller-5bddd4b946-7c5ls\" (UID: \"a8825499-558f-4b31-b09b-0dcc92662f65\") " pod="metallb-system/controller-5bddd4b946-7c5ls" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.394605 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-5bddd4b946-7c5ls" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.630775 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-metrics-certs\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.634602 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd-metrics-certs\") pod \"frr-k8s-v2lll\" (UID: \"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd\") " pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.719500 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7784b6fcf-5bsq9"] Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.732145 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ea01c7df-c5aa-4c86-800b-b6aebafeb8fd-metrics-certs\") pod \"speaker-4jwwj\" (UID: \"ea01c7df-c5aa-4c86-800b-b6aebafeb8fd\") " pod="metallb-system/speaker-4jwwj" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.732467 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ea01c7df-c5aa-4c86-800b-b6aebafeb8fd-memberlist\") pod \"speaker-4jwwj\" (UID: \"ea01c7df-c5aa-4c86-800b-b6aebafeb8fd\") " pod="metallb-system/speaker-4jwwj" Dec 27 07:59:07 crc kubenswrapper[4934]: E1227 07:59:07.732623 4934 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 27 07:59:07 crc kubenswrapper[4934]: E1227 07:59:07.732679 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ea01c7df-c5aa-4c86-800b-b6aebafeb8fd-memberlist podName:ea01c7df-c5aa-4c86-800b-b6aebafeb8fd nodeName:}" failed. No retries permitted until 2025-12-27 07:59:08.732664193 +0000 UTC m=+1009.553104787 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/ea01c7df-c5aa-4c86-800b-b6aebafeb8fd-memberlist") pod "speaker-4jwwj" (UID: "ea01c7df-c5aa-4c86-800b-b6aebafeb8fd") : secret "metallb-memberlist" not found Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.737462 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ea01c7df-c5aa-4c86-800b-b6aebafeb8fd-metrics-certs\") pod \"speaker-4jwwj\" (UID: \"ea01c7df-c5aa-4c86-800b-b6aebafeb8fd\") " pod="metallb-system/speaker-4jwwj" Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.867836 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-5bddd4b946-7c5ls"] Dec 27 07:59:07 crc kubenswrapper[4934]: I1227 07:59:07.869584 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:07 crc kubenswrapper[4934]: W1227 07:59:07.872067 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8825499_558f_4b31_b09b_0dcc92662f65.slice/crio-648011d0c19fba169f9bb1d5fe0ffb11988a5316a5f6946fe298954b6f857e6f WatchSource:0}: Error finding container 648011d0c19fba169f9bb1d5fe0ffb11988a5316a5f6946fe298954b6f857e6f: Status 404 returned error can't find the container with id 648011d0c19fba169f9bb1d5fe0ffb11988a5316a5f6946fe298954b6f857e6f Dec 27 07:59:08 crc kubenswrapper[4934]: I1227 07:59:08.697779 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2lll" event={"ID":"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd","Type":"ContainerStarted","Data":"5e37bb7c25cb647b5f45be8a6370f8fd2ca1a33a78c564844326227397a6ef4e"} Dec 27 07:59:08 crc kubenswrapper[4934]: I1227 07:59:08.699033 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5bsq9" event={"ID":"0134c36f-c43b-4830-848a-3a8690957ee1","Type":"ContainerStarted","Data":"6d3a9317b68f5b9968fe82c7ed299eb3924ead9a93a70400f790f4cfc1ca692b"} Dec 27 07:59:08 crc kubenswrapper[4934]: I1227 07:59:08.701146 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-7c5ls" event={"ID":"a8825499-558f-4b31-b09b-0dcc92662f65","Type":"ContainerStarted","Data":"5b5c5879c0ba0ed6fb65d21719312b133d13bc08312c4bf3310f2485e5de2154"} Dec 27 07:59:08 crc kubenswrapper[4934]: I1227 07:59:08.701197 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-7c5ls" event={"ID":"a8825499-558f-4b31-b09b-0dcc92662f65","Type":"ContainerStarted","Data":"e5c694ae386e0a214260677899c88bc6eaa638ec88f3199e925722a091724206"} Dec 27 07:59:08 crc kubenswrapper[4934]: I1227 07:59:08.701213 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-7c5ls" event={"ID":"a8825499-558f-4b31-b09b-0dcc92662f65","Type":"ContainerStarted","Data":"648011d0c19fba169f9bb1d5fe0ffb11988a5316a5f6946fe298954b6f857e6f"} Dec 27 07:59:08 crc kubenswrapper[4934]: I1227 07:59:08.701321 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5bddd4b946-7c5ls" Dec 27 07:59:08 crc kubenswrapper[4934]: I1227 07:59:08.720877 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-5bddd4b946-7c5ls" podStartSLOduration=1.7208616810000001 podStartE2EDuration="1.720861681s" podCreationTimestamp="2025-12-27 07:59:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:59:08.716849979 +0000 UTC m=+1009.537290583" watchObservedRunningTime="2025-12-27 07:59:08.720861681 +0000 UTC m=+1009.541302275" Dec 27 07:59:08 crc kubenswrapper[4934]: I1227 07:59:08.751749 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ea01c7df-c5aa-4c86-800b-b6aebafeb8fd-memberlist\") pod \"speaker-4jwwj\" (UID: \"ea01c7df-c5aa-4c86-800b-b6aebafeb8fd\") " pod="metallb-system/speaker-4jwwj" Dec 27 07:59:08 crc kubenswrapper[4934]: I1227 07:59:08.756558 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ea01c7df-c5aa-4c86-800b-b6aebafeb8fd-memberlist\") pod \"speaker-4jwwj\" (UID: \"ea01c7df-c5aa-4c86-800b-b6aebafeb8fd\") " pod="metallb-system/speaker-4jwwj" Dec 27 07:59:08 crc kubenswrapper[4934]: I1227 07:59:08.872582 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-4jwwj" Dec 27 07:59:08 crc kubenswrapper[4934]: W1227 07:59:08.913934 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea01c7df_c5aa_4c86_800b_b6aebafeb8fd.slice/crio-87032489e4b4a5fab0c4d68e997c55bcfcca4fbe3f84568655f484a2df0316eb WatchSource:0}: Error finding container 87032489e4b4a5fab0c4d68e997c55bcfcca4fbe3f84568655f484a2df0316eb: Status 404 returned error can't find the container with id 87032489e4b4a5fab0c4d68e997c55bcfcca4fbe3f84568655f484a2df0316eb Dec 27 07:59:09 crc kubenswrapper[4934]: I1227 07:59:09.715011 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-4jwwj" event={"ID":"ea01c7df-c5aa-4c86-800b-b6aebafeb8fd","Type":"ContainerStarted","Data":"6082d2fb77aa6da7250a56f3df3ec616ef86844616316586bfaccea202d6cf17"} Dec 27 07:59:09 crc kubenswrapper[4934]: I1227 07:59:09.715349 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-4jwwj" event={"ID":"ea01c7df-c5aa-4c86-800b-b6aebafeb8fd","Type":"ContainerStarted","Data":"4fdaa4d99182e572e8fa5c0576fa8376e9a7081a1b2c57612142ae52ae3eff82"} Dec 27 07:59:09 crc kubenswrapper[4934]: I1227 07:59:09.715362 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-4jwwj" event={"ID":"ea01c7df-c5aa-4c86-800b-b6aebafeb8fd","Type":"ContainerStarted","Data":"87032489e4b4a5fab0c4d68e997c55bcfcca4fbe3f84568655f484a2df0316eb"} Dec 27 07:59:09 crc kubenswrapper[4934]: I1227 07:59:09.715512 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-4jwwj" Dec 27 07:59:09 crc kubenswrapper[4934]: I1227 07:59:09.730894 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-4jwwj" podStartSLOduration=2.7308752800000002 podStartE2EDuration="2.73087528s" podCreationTimestamp="2025-12-27 07:59:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 07:59:09.728553042 +0000 UTC m=+1010.548993656" watchObservedRunningTime="2025-12-27 07:59:09.73087528 +0000 UTC m=+1010.551315874" Dec 27 07:59:15 crc kubenswrapper[4934]: I1227 07:59:15.809388 4934 generic.go:334] "Generic (PLEG): container finished" podID="c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd" containerID="b3c199073009086b7f9c8d82deb00ebeb5440ba5793bfd3e77f4e2ee6a905b62" exitCode=0 Dec 27 07:59:15 crc kubenswrapper[4934]: I1227 07:59:15.809473 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2lll" event={"ID":"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd","Type":"ContainerDied","Data":"b3c199073009086b7f9c8d82deb00ebeb5440ba5793bfd3e77f4e2ee6a905b62"} Dec 27 07:59:15 crc kubenswrapper[4934]: I1227 07:59:15.813989 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5bsq9" event={"ID":"0134c36f-c43b-4830-848a-3a8690957ee1","Type":"ContainerStarted","Data":"afa23a439758adce15ad845e8c934b8b67c26b78f0b2c26b7f406620dd32c6c4"} Dec 27 07:59:15 crc kubenswrapper[4934]: I1227 07:59:15.814580 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5bsq9" Dec 27 07:59:15 crc kubenswrapper[4934]: I1227 07:59:15.895369 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5bsq9" podStartSLOduration=2.390679885 podStartE2EDuration="9.895335288s" podCreationTimestamp="2025-12-27 07:59:06 +0000 UTC" firstStartedPulling="2025-12-27 07:59:07.734685794 +0000 UTC m=+1008.555126398" lastFinishedPulling="2025-12-27 07:59:15.239341157 +0000 UTC m=+1016.059781801" observedRunningTime="2025-12-27 07:59:15.884253488 +0000 UTC m=+1016.704694112" watchObservedRunningTime="2025-12-27 07:59:15.895335288 +0000 UTC m=+1016.715775912" Dec 27 07:59:16 crc kubenswrapper[4934]: I1227 07:59:16.826379 4934 generic.go:334] "Generic (PLEG): container finished" podID="c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd" containerID="de65541c8f24aed6c9ee3eecf784c1f20664e7586bc3df88b8a619afb4286d22" exitCode=0 Dec 27 07:59:16 crc kubenswrapper[4934]: I1227 07:59:16.826603 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2lll" event={"ID":"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd","Type":"ContainerDied","Data":"de65541c8f24aed6c9ee3eecf784c1f20664e7586bc3df88b8a619afb4286d22"} Dec 27 07:59:17 crc kubenswrapper[4934]: I1227 07:59:17.837016 4934 generic.go:334] "Generic (PLEG): container finished" podID="c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd" containerID="e8d89c59f59e8005625a9d08ea30cc5062bbc305153f73ca08aac6caccfbdb69" exitCode=0 Dec 27 07:59:17 crc kubenswrapper[4934]: I1227 07:59:17.837136 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2lll" event={"ID":"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd","Type":"ContainerDied","Data":"e8d89c59f59e8005625a9d08ea30cc5062bbc305153f73ca08aac6caccfbdb69"} Dec 27 07:59:18 crc kubenswrapper[4934]: I1227 07:59:18.849318 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2lll" event={"ID":"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd","Type":"ContainerStarted","Data":"52782e5d8a5dc30cd8cee49bd73f9845c4e84e0c8d6d47bc490e378d1f6ebbb5"} Dec 27 07:59:18 crc kubenswrapper[4934]: I1227 07:59:18.849711 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2lll" event={"ID":"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd","Type":"ContainerStarted","Data":"55c1ffe1f08ad37eb20cc55bf2939230fef1fa017098d2cd169a0a4686f8a216"} Dec 27 07:59:18 crc kubenswrapper[4934]: I1227 07:59:18.849735 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2lll" event={"ID":"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd","Type":"ContainerStarted","Data":"0ef9b620b193a8fcd11fab4e244104037e9a038cd1fdc43daa69d30cf4b9ad84"} Dec 27 07:59:18 crc kubenswrapper[4934]: I1227 07:59:18.849751 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2lll" event={"ID":"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd","Type":"ContainerStarted","Data":"a23c34a111bcac9a8c777af6384e620fb75b23d9f584e874f91638de1d4a1977"} Dec 27 07:59:18 crc kubenswrapper[4934]: I1227 07:59:18.849789 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2lll" event={"ID":"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd","Type":"ContainerStarted","Data":"eb76c6df7ecc4237bb03fa19dd8352961a29e61d9c0d539a252ee3be2aac3fce"} Dec 27 07:59:19 crc kubenswrapper[4934]: I1227 07:59:19.863768 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2lll" event={"ID":"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd","Type":"ContainerStarted","Data":"ded2286bdfa3a5a9a07cd74d23616128ad1917729e3f65fa7322ae65d4b9990f"} Dec 27 07:59:19 crc kubenswrapper[4934]: I1227 07:59:19.864441 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:19 crc kubenswrapper[4934]: I1227 07:59:19.902836 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-v2lll" podStartSLOduration=6.658945279 podStartE2EDuration="13.902818205s" podCreationTimestamp="2025-12-27 07:59:06 +0000 UTC" firstStartedPulling="2025-12-27 07:59:08.019242104 +0000 UTC m=+1008.839682718" lastFinishedPulling="2025-12-27 07:59:15.26311501 +0000 UTC m=+1016.083555644" observedRunningTime="2025-12-27 07:59:19.898878625 +0000 UTC m=+1020.719319259" watchObservedRunningTime="2025-12-27 07:59:19.902818205 +0000 UTC m=+1020.723258799" Dec 27 07:59:22 crc kubenswrapper[4934]: I1227 07:59:22.870861 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:22 crc kubenswrapper[4934]: I1227 07:59:22.917396 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:27 crc kubenswrapper[4934]: I1227 07:59:27.300217 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5bsq9" Dec 27 07:59:27 crc kubenswrapper[4934]: I1227 07:59:27.403195 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5bddd4b946-7c5ls" Dec 27 07:59:28 crc kubenswrapper[4934]: I1227 07:59:28.882248 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-4jwwj" Dec 27 07:59:32 crc kubenswrapper[4934]: I1227 07:59:32.010184 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-g8xgs"] Dec 27 07:59:32 crc kubenswrapper[4934]: I1227 07:59:32.012153 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-g8xgs" Dec 27 07:59:32 crc kubenswrapper[4934]: I1227 07:59:32.014318 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-55b8j" Dec 27 07:59:32 crc kubenswrapper[4934]: I1227 07:59:32.015024 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 27 07:59:32 crc kubenswrapper[4934]: I1227 07:59:32.015303 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 27 07:59:32 crc kubenswrapper[4934]: I1227 07:59:32.044637 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-g8xgs"] Dec 27 07:59:32 crc kubenswrapper[4934]: I1227 07:59:32.112554 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj5cn\" (UniqueName: \"kubernetes.io/projected/6a92cb45-7d29-472c-a674-f0bf9faad41a-kube-api-access-lj5cn\") pod \"openstack-operator-index-g8xgs\" (UID: \"6a92cb45-7d29-472c-a674-f0bf9faad41a\") " pod="openstack-operators/openstack-operator-index-g8xgs" Dec 27 07:59:32 crc kubenswrapper[4934]: I1227 07:59:32.214188 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lj5cn\" (UniqueName: \"kubernetes.io/projected/6a92cb45-7d29-472c-a674-f0bf9faad41a-kube-api-access-lj5cn\") pod \"openstack-operator-index-g8xgs\" (UID: \"6a92cb45-7d29-472c-a674-f0bf9faad41a\") " pod="openstack-operators/openstack-operator-index-g8xgs" Dec 27 07:59:32 crc kubenswrapper[4934]: I1227 07:59:32.229963 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj5cn\" (UniqueName: \"kubernetes.io/projected/6a92cb45-7d29-472c-a674-f0bf9faad41a-kube-api-access-lj5cn\") pod \"openstack-operator-index-g8xgs\" (UID: \"6a92cb45-7d29-472c-a674-f0bf9faad41a\") " pod="openstack-operators/openstack-operator-index-g8xgs" Dec 27 07:59:32 crc kubenswrapper[4934]: I1227 07:59:32.331599 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-g8xgs" Dec 27 07:59:32 crc kubenswrapper[4934]: I1227 07:59:32.842325 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-g8xgs"] Dec 27 07:59:32 crc kubenswrapper[4934]: W1227 07:59:32.852708 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a92cb45_7d29_472c_a674_f0bf9faad41a.slice/crio-ea63e48764f854ed991ac4194214b8e05987fb260c781601fc40f0706d90a8ee WatchSource:0}: Error finding container ea63e48764f854ed991ac4194214b8e05987fb260c781601fc40f0706d90a8ee: Status 404 returned error can't find the container with id ea63e48764f854ed991ac4194214b8e05987fb260c781601fc40f0706d90a8ee Dec 27 07:59:32 crc kubenswrapper[4934]: I1227 07:59:32.995176 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-g8xgs" event={"ID":"6a92cb45-7d29-472c-a674-f0bf9faad41a","Type":"ContainerStarted","Data":"ea63e48764f854ed991ac4194214b8e05987fb260c781601fc40f0706d90a8ee"} Dec 27 07:59:35 crc kubenswrapper[4934]: I1227 07:59:35.016056 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-g8xgs" event={"ID":"6a92cb45-7d29-472c-a674-f0bf9faad41a","Type":"ContainerStarted","Data":"c91e0e6276d417244e5c15beb87c3557ec8a54cec7c60732ed297cf8a510d688"} Dec 27 07:59:35 crc kubenswrapper[4934]: I1227 07:59:35.040429 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-g8xgs" podStartSLOduration=2.362797346 podStartE2EDuration="4.040397861s" podCreationTimestamp="2025-12-27 07:59:31 +0000 UTC" firstStartedPulling="2025-12-27 07:59:32.85653606 +0000 UTC m=+1033.676976694" lastFinishedPulling="2025-12-27 07:59:34.534136605 +0000 UTC m=+1035.354577209" observedRunningTime="2025-12-27 07:59:35.03520708 +0000 UTC m=+1035.855647764" watchObservedRunningTime="2025-12-27 07:59:35.040397861 +0000 UTC m=+1035.860838495" Dec 27 07:59:35 crc kubenswrapper[4934]: I1227 07:59:35.379502 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-g8xgs"] Dec 27 07:59:35 crc kubenswrapper[4934]: I1227 07:59:35.991898 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-v7q5s"] Dec 27 07:59:35 crc kubenswrapper[4934]: I1227 07:59:35.993231 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-v7q5s" Dec 27 07:59:36 crc kubenswrapper[4934]: I1227 07:59:36.003949 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-v7q5s"] Dec 27 07:59:36 crc kubenswrapper[4934]: I1227 07:59:36.094496 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5snw\" (UniqueName: \"kubernetes.io/projected/a41afe73-59e9-4cbe-b2e7-1f755767daf5-kube-api-access-m5snw\") pod \"openstack-operator-index-v7q5s\" (UID: \"a41afe73-59e9-4cbe-b2e7-1f755767daf5\") " pod="openstack-operators/openstack-operator-index-v7q5s" Dec 27 07:59:36 crc kubenswrapper[4934]: I1227 07:59:36.196466 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5snw\" (UniqueName: \"kubernetes.io/projected/a41afe73-59e9-4cbe-b2e7-1f755767daf5-kube-api-access-m5snw\") pod \"openstack-operator-index-v7q5s\" (UID: \"a41afe73-59e9-4cbe-b2e7-1f755767daf5\") " pod="openstack-operators/openstack-operator-index-v7q5s" Dec 27 07:59:36 crc kubenswrapper[4934]: I1227 07:59:36.229231 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5snw\" (UniqueName: \"kubernetes.io/projected/a41afe73-59e9-4cbe-b2e7-1f755767daf5-kube-api-access-m5snw\") pod \"openstack-operator-index-v7q5s\" (UID: \"a41afe73-59e9-4cbe-b2e7-1f755767daf5\") " pod="openstack-operators/openstack-operator-index-v7q5s" Dec 27 07:59:36 crc kubenswrapper[4934]: I1227 07:59:36.321566 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-v7q5s" Dec 27 07:59:36 crc kubenswrapper[4934]: I1227 07:59:36.868934 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-v7q5s"] Dec 27 07:59:37 crc kubenswrapper[4934]: I1227 07:59:37.034712 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-v7q5s" event={"ID":"a41afe73-59e9-4cbe-b2e7-1f755767daf5","Type":"ContainerStarted","Data":"fa3bc91505ccb81d4a3d8e61cf880c906a6f798b88917019855cc29a17977740"} Dec 27 07:59:37 crc kubenswrapper[4934]: I1227 07:59:37.034853 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-g8xgs" podUID="6a92cb45-7d29-472c-a674-f0bf9faad41a" containerName="registry-server" containerID="cri-o://c91e0e6276d417244e5c15beb87c3557ec8a54cec7c60732ed297cf8a510d688" gracePeriod=2 Dec 27 07:59:37 crc kubenswrapper[4934]: I1227 07:59:37.590502 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-g8xgs" Dec 27 07:59:37 crc kubenswrapper[4934]: I1227 07:59:37.722422 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lj5cn\" (UniqueName: \"kubernetes.io/projected/6a92cb45-7d29-472c-a674-f0bf9faad41a-kube-api-access-lj5cn\") pod \"6a92cb45-7d29-472c-a674-f0bf9faad41a\" (UID: \"6a92cb45-7d29-472c-a674-f0bf9faad41a\") " Dec 27 07:59:37 crc kubenswrapper[4934]: I1227 07:59:37.732760 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a92cb45-7d29-472c-a674-f0bf9faad41a-kube-api-access-lj5cn" (OuterVolumeSpecName: "kube-api-access-lj5cn") pod "6a92cb45-7d29-472c-a674-f0bf9faad41a" (UID: "6a92cb45-7d29-472c-a674-f0bf9faad41a"). InnerVolumeSpecName "kube-api-access-lj5cn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 07:59:37 crc kubenswrapper[4934]: I1227 07:59:37.823971 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lj5cn\" (UniqueName: \"kubernetes.io/projected/6a92cb45-7d29-472c-a674-f0bf9faad41a-kube-api-access-lj5cn\") on node \"crc\" DevicePath \"\"" Dec 27 07:59:37 crc kubenswrapper[4934]: I1227 07:59:37.875583 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-v2lll" Dec 27 07:59:38 crc kubenswrapper[4934]: I1227 07:59:38.043197 4934 generic.go:334] "Generic (PLEG): container finished" podID="6a92cb45-7d29-472c-a674-f0bf9faad41a" containerID="c91e0e6276d417244e5c15beb87c3557ec8a54cec7c60732ed297cf8a510d688" exitCode=0 Dec 27 07:59:38 crc kubenswrapper[4934]: I1227 07:59:38.043255 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-g8xgs" Dec 27 07:59:38 crc kubenswrapper[4934]: I1227 07:59:38.043266 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-g8xgs" event={"ID":"6a92cb45-7d29-472c-a674-f0bf9faad41a","Type":"ContainerDied","Data":"c91e0e6276d417244e5c15beb87c3557ec8a54cec7c60732ed297cf8a510d688"} Dec 27 07:59:38 crc kubenswrapper[4934]: I1227 07:59:38.043390 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-g8xgs" event={"ID":"6a92cb45-7d29-472c-a674-f0bf9faad41a","Type":"ContainerDied","Data":"ea63e48764f854ed991ac4194214b8e05987fb260c781601fc40f0706d90a8ee"} Dec 27 07:59:38 crc kubenswrapper[4934]: I1227 07:59:38.043426 4934 scope.go:117] "RemoveContainer" containerID="c91e0e6276d417244e5c15beb87c3557ec8a54cec7c60732ed297cf8a510d688" Dec 27 07:59:38 crc kubenswrapper[4934]: I1227 07:59:38.044627 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-v7q5s" event={"ID":"a41afe73-59e9-4cbe-b2e7-1f755767daf5","Type":"ContainerStarted","Data":"c4d90a64d6b43dc3fc3e25234dfa273b006d63534bedde5f29edbdb6e5764265"} Dec 27 07:59:38 crc kubenswrapper[4934]: I1227 07:59:38.070979 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-v7q5s" podStartSLOduration=2.984928788 podStartE2EDuration="3.070939447s" podCreationTimestamp="2025-12-27 07:59:35 +0000 UTC" firstStartedPulling="2025-12-27 07:59:36.881737916 +0000 UTC m=+1037.702178520" lastFinishedPulling="2025-12-27 07:59:36.967748575 +0000 UTC m=+1037.788189179" observedRunningTime="2025-12-27 07:59:38.068217338 +0000 UTC m=+1038.888657942" watchObservedRunningTime="2025-12-27 07:59:38.070939447 +0000 UTC m=+1038.891380041" Dec 27 07:59:38 crc kubenswrapper[4934]: I1227 07:59:38.091715 4934 scope.go:117] "RemoveContainer" containerID="c91e0e6276d417244e5c15beb87c3557ec8a54cec7c60732ed297cf8a510d688" Dec 27 07:59:38 crc kubenswrapper[4934]: I1227 07:59:38.096939 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-g8xgs"] Dec 27 07:59:38 crc kubenswrapper[4934]: E1227 07:59:38.099451 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c91e0e6276d417244e5c15beb87c3557ec8a54cec7c60732ed297cf8a510d688\": container with ID starting with c91e0e6276d417244e5c15beb87c3557ec8a54cec7c60732ed297cf8a510d688 not found: ID does not exist" containerID="c91e0e6276d417244e5c15beb87c3557ec8a54cec7c60732ed297cf8a510d688" Dec 27 07:59:38 crc kubenswrapper[4934]: I1227 07:59:38.099574 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c91e0e6276d417244e5c15beb87c3557ec8a54cec7c60732ed297cf8a510d688"} err="failed to get container status \"c91e0e6276d417244e5c15beb87c3557ec8a54cec7c60732ed297cf8a510d688\": rpc error: code = NotFound desc = could not find container \"c91e0e6276d417244e5c15beb87c3557ec8a54cec7c60732ed297cf8a510d688\": container with ID starting with c91e0e6276d417244e5c15beb87c3557ec8a54cec7c60732ed297cf8a510d688 not found: ID does not exist" Dec 27 07:59:38 crc kubenswrapper[4934]: I1227 07:59:38.106164 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-g8xgs"] Dec 27 07:59:39 crc kubenswrapper[4934]: I1227 07:59:39.484001 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a92cb45-7d29-472c-a674-f0bf9faad41a" path="/var/lib/kubelet/pods/6a92cb45-7d29-472c-a674-f0bf9faad41a/volumes" Dec 27 07:59:46 crc kubenswrapper[4934]: I1227 07:59:46.322277 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-v7q5s" Dec 27 07:59:46 crc kubenswrapper[4934]: I1227 07:59:46.323272 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-v7q5s" Dec 27 07:59:46 crc kubenswrapper[4934]: I1227 07:59:46.371068 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-v7q5s" Dec 27 07:59:47 crc kubenswrapper[4934]: I1227 07:59:47.173310 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-v7q5s" Dec 27 07:59:55 crc kubenswrapper[4934]: I1227 07:59:55.204016 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv"] Dec 27 07:59:55 crc kubenswrapper[4934]: E1227 07:59:55.205097 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a92cb45-7d29-472c-a674-f0bf9faad41a" containerName="registry-server" Dec 27 07:59:55 crc kubenswrapper[4934]: I1227 07:59:55.205113 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a92cb45-7d29-472c-a674-f0bf9faad41a" containerName="registry-server" Dec 27 07:59:55 crc kubenswrapper[4934]: I1227 07:59:55.205278 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a92cb45-7d29-472c-a674-f0bf9faad41a" containerName="registry-server" Dec 27 07:59:55 crc kubenswrapper[4934]: I1227 07:59:55.206359 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv" Dec 27 07:59:55 crc kubenswrapper[4934]: I1227 07:59:55.209404 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-8kx9h" Dec 27 07:59:55 crc kubenswrapper[4934]: I1227 07:59:55.220139 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv"] Dec 27 07:59:55 crc kubenswrapper[4934]: I1227 07:59:55.375679 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bms6c\" (UniqueName: \"kubernetes.io/projected/3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4-kube-api-access-bms6c\") pod \"ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv\" (UID: \"3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4\") " pod="openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv" Dec 27 07:59:55 crc kubenswrapper[4934]: I1227 07:59:55.375734 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4-bundle\") pod \"ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv\" (UID: \"3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4\") " pod="openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv" Dec 27 07:59:55 crc kubenswrapper[4934]: I1227 07:59:55.375762 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4-util\") pod \"ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv\" (UID: \"3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4\") " pod="openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv" Dec 27 07:59:55 crc kubenswrapper[4934]: I1227 07:59:55.477339 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4-bundle\") pod \"ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv\" (UID: \"3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4\") " pod="openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv" Dec 27 07:59:55 crc kubenswrapper[4934]: I1227 07:59:55.477401 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4-util\") pod \"ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv\" (UID: \"3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4\") " pod="openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv" Dec 27 07:59:55 crc kubenswrapper[4934]: I1227 07:59:55.477556 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bms6c\" (UniqueName: \"kubernetes.io/projected/3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4-kube-api-access-bms6c\") pod \"ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv\" (UID: \"3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4\") " pod="openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv" Dec 27 07:59:55 crc kubenswrapper[4934]: I1227 07:59:55.478032 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4-bundle\") pod \"ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv\" (UID: \"3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4\") " pod="openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv" Dec 27 07:59:55 crc kubenswrapper[4934]: I1227 07:59:55.478231 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4-util\") pod \"ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv\" (UID: \"3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4\") " pod="openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv" Dec 27 07:59:55 crc kubenswrapper[4934]: I1227 07:59:55.507236 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bms6c\" (UniqueName: \"kubernetes.io/projected/3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4-kube-api-access-bms6c\") pod \"ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv\" (UID: \"3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4\") " pod="openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv" Dec 27 07:59:55 crc kubenswrapper[4934]: I1227 07:59:55.523666 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv" Dec 27 07:59:55 crc kubenswrapper[4934]: I1227 07:59:55.961924 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv"] Dec 27 07:59:55 crc kubenswrapper[4934]: W1227 07:59:55.994345 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3fb5ccd1_1f57_4b3b_97d2_97af95b73ec4.slice/crio-626453714b5d0f8c13494a15a33fba5231c079f6f158b197eb12dcaa0ff88009 WatchSource:0}: Error finding container 626453714b5d0f8c13494a15a33fba5231c079f6f158b197eb12dcaa0ff88009: Status 404 returned error can't find the container with id 626453714b5d0f8c13494a15a33fba5231c079f6f158b197eb12dcaa0ff88009 Dec 27 07:59:56 crc kubenswrapper[4934]: I1227 07:59:56.228442 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv" event={"ID":"3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4","Type":"ContainerStarted","Data":"4fa473d075cc1205555df3732f15a93a7a9160d0413704fd0cdd43bc589575b4"} Dec 27 07:59:56 crc kubenswrapper[4934]: I1227 07:59:56.228875 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv" event={"ID":"3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4","Type":"ContainerStarted","Data":"626453714b5d0f8c13494a15a33fba5231c079f6f158b197eb12dcaa0ff88009"} Dec 27 07:59:57 crc kubenswrapper[4934]: I1227 07:59:57.244196 4934 generic.go:334] "Generic (PLEG): container finished" podID="3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4" containerID="4fa473d075cc1205555df3732f15a93a7a9160d0413704fd0cdd43bc589575b4" exitCode=0 Dec 27 07:59:57 crc kubenswrapper[4934]: I1227 07:59:57.244331 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv" event={"ID":"3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4","Type":"ContainerDied","Data":"4fa473d075cc1205555df3732f15a93a7a9160d0413704fd0cdd43bc589575b4"} Dec 27 07:59:58 crc kubenswrapper[4934]: I1227 07:59:58.261767 4934 generic.go:334] "Generic (PLEG): container finished" podID="3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4" containerID="1ce44d872bc3127df7dc5066bcc79a10dda8aad229223d0c90bcb05090deda96" exitCode=0 Dec 27 07:59:58 crc kubenswrapper[4934]: I1227 07:59:58.261870 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv" event={"ID":"3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4","Type":"ContainerDied","Data":"1ce44d872bc3127df7dc5066bcc79a10dda8aad229223d0c90bcb05090deda96"} Dec 27 07:59:59 crc kubenswrapper[4934]: I1227 07:59:59.273585 4934 generic.go:334] "Generic (PLEG): container finished" podID="3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4" containerID="fdcf0cf11974b5da7c9d4a51b81ddff1e97258e74dc78ccaa7d55844b58f2020" exitCode=0 Dec 27 07:59:59 crc kubenswrapper[4934]: I1227 07:59:59.273644 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv" event={"ID":"3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4","Type":"ContainerDied","Data":"fdcf0cf11974b5da7c9d4a51b81ddff1e97258e74dc78ccaa7d55844b58f2020"} Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.194756 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg"] Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.197252 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg" Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.200276 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.200590 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.209215 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg"] Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.359963 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5885f5e1-fac5-4980-8501-f882f612fa04-config-volume\") pod \"collect-profiles-29447040-pw7gg\" (UID: \"5885f5e1-fac5-4980-8501-f882f612fa04\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg" Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.360186 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crhk4\" (UniqueName: \"kubernetes.io/projected/5885f5e1-fac5-4980-8501-f882f612fa04-kube-api-access-crhk4\") pod \"collect-profiles-29447040-pw7gg\" (UID: \"5885f5e1-fac5-4980-8501-f882f612fa04\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg" Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.360474 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5885f5e1-fac5-4980-8501-f882f612fa04-secret-volume\") pod \"collect-profiles-29447040-pw7gg\" (UID: \"5885f5e1-fac5-4980-8501-f882f612fa04\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg" Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.462371 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5885f5e1-fac5-4980-8501-f882f612fa04-config-volume\") pod \"collect-profiles-29447040-pw7gg\" (UID: \"5885f5e1-fac5-4980-8501-f882f612fa04\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg" Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.462457 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crhk4\" (UniqueName: \"kubernetes.io/projected/5885f5e1-fac5-4980-8501-f882f612fa04-kube-api-access-crhk4\") pod \"collect-profiles-29447040-pw7gg\" (UID: \"5885f5e1-fac5-4980-8501-f882f612fa04\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg" Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.462562 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5885f5e1-fac5-4980-8501-f882f612fa04-secret-volume\") pod \"collect-profiles-29447040-pw7gg\" (UID: \"5885f5e1-fac5-4980-8501-f882f612fa04\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg" Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.463660 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5885f5e1-fac5-4980-8501-f882f612fa04-config-volume\") pod \"collect-profiles-29447040-pw7gg\" (UID: \"5885f5e1-fac5-4980-8501-f882f612fa04\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg" Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.488752 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5885f5e1-fac5-4980-8501-f882f612fa04-secret-volume\") pod \"collect-profiles-29447040-pw7gg\" (UID: \"5885f5e1-fac5-4980-8501-f882f612fa04\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg" Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.492495 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crhk4\" (UniqueName: \"kubernetes.io/projected/5885f5e1-fac5-4980-8501-f882f612fa04-kube-api-access-crhk4\") pod \"collect-profiles-29447040-pw7gg\" (UID: \"5885f5e1-fac5-4980-8501-f882f612fa04\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg" Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.523102 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg" Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.639716 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv" Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.769499 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4-util\") pod \"3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4\" (UID: \"3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4\") " Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.769621 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4-bundle\") pod \"3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4\" (UID: \"3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4\") " Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.769698 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bms6c\" (UniqueName: \"kubernetes.io/projected/3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4-kube-api-access-bms6c\") pod \"3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4\" (UID: \"3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4\") " Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.770998 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4-bundle" (OuterVolumeSpecName: "bundle") pod "3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4" (UID: "3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.777322 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4-kube-api-access-bms6c" (OuterVolumeSpecName: "kube-api-access-bms6c") pod "3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4" (UID: "3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4"). InnerVolumeSpecName "kube-api-access-bms6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.793600 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4-util" (OuterVolumeSpecName: "util") pod "3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4" (UID: "3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.872480 4934 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4-util\") on node \"crc\" DevicePath \"\"" Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.872545 4934 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.872571 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bms6c\" (UniqueName: \"kubernetes.io/projected/3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4-kube-api-access-bms6c\") on node \"crc\" DevicePath \"\"" Dec 27 08:00:00 crc kubenswrapper[4934]: I1227 08:00:00.994971 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg"] Dec 27 08:00:01 crc kubenswrapper[4934]: W1227 08:00:01.000673 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5885f5e1_fac5_4980_8501_f882f612fa04.slice/crio-b3a492eb0d415687a6b4ac86ef95499ffabe6225a357bd00b0cd7e9f6cda0496 WatchSource:0}: Error finding container b3a492eb0d415687a6b4ac86ef95499ffabe6225a357bd00b0cd7e9f6cda0496: Status 404 returned error can't find the container with id b3a492eb0d415687a6b4ac86ef95499ffabe6225a357bd00b0cd7e9f6cda0496 Dec 27 08:00:01 crc kubenswrapper[4934]: I1227 08:00:01.291468 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg" event={"ID":"5885f5e1-fac5-4980-8501-f882f612fa04","Type":"ContainerStarted","Data":"e4c25105ac81a169026cb2bd288bf2ce3bb8524168682d6cdab900409c04d0f3"} Dec 27 08:00:01 crc kubenswrapper[4934]: I1227 08:00:01.291803 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg" event={"ID":"5885f5e1-fac5-4980-8501-f882f612fa04","Type":"ContainerStarted","Data":"b3a492eb0d415687a6b4ac86ef95499ffabe6225a357bd00b0cd7e9f6cda0496"} Dec 27 08:00:01 crc kubenswrapper[4934]: I1227 08:00:01.295743 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv" event={"ID":"3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4","Type":"ContainerDied","Data":"626453714b5d0f8c13494a15a33fba5231c079f6f158b197eb12dcaa0ff88009"} Dec 27 08:00:01 crc kubenswrapper[4934]: I1227 08:00:01.295781 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="626453714b5d0f8c13494a15a33fba5231c079f6f158b197eb12dcaa0ff88009" Dec 27 08:00:01 crc kubenswrapper[4934]: I1227 08:00:01.295798 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv" Dec 27 08:00:01 crc kubenswrapper[4934]: I1227 08:00:01.316244 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg" podStartSLOduration=1.316228756 podStartE2EDuration="1.316228756s" podCreationTimestamp="2025-12-27 08:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:00:01.312386059 +0000 UTC m=+1062.132826673" watchObservedRunningTime="2025-12-27 08:00:01.316228756 +0000 UTC m=+1062.136669350" Dec 27 08:00:02 crc kubenswrapper[4934]: I1227 08:00:02.309522 4934 generic.go:334] "Generic (PLEG): container finished" podID="5885f5e1-fac5-4980-8501-f882f612fa04" containerID="e4c25105ac81a169026cb2bd288bf2ce3bb8524168682d6cdab900409c04d0f3" exitCode=0 Dec 27 08:00:02 crc kubenswrapper[4934]: I1227 08:00:02.309570 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg" event={"ID":"5885f5e1-fac5-4980-8501-f882f612fa04","Type":"ContainerDied","Data":"e4c25105ac81a169026cb2bd288bf2ce3bb8524168682d6cdab900409c04d0f3"} Dec 27 08:00:03 crc kubenswrapper[4934]: I1227 08:00:03.783585 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg" Dec 27 08:00:03 crc kubenswrapper[4934]: I1227 08:00:03.820620 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5885f5e1-fac5-4980-8501-f882f612fa04-secret-volume\") pod \"5885f5e1-fac5-4980-8501-f882f612fa04\" (UID: \"5885f5e1-fac5-4980-8501-f882f612fa04\") " Dec 27 08:00:03 crc kubenswrapper[4934]: I1227 08:00:03.820691 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crhk4\" (UniqueName: \"kubernetes.io/projected/5885f5e1-fac5-4980-8501-f882f612fa04-kube-api-access-crhk4\") pod \"5885f5e1-fac5-4980-8501-f882f612fa04\" (UID: \"5885f5e1-fac5-4980-8501-f882f612fa04\") " Dec 27 08:00:03 crc kubenswrapper[4934]: I1227 08:00:03.820927 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5885f5e1-fac5-4980-8501-f882f612fa04-config-volume\") pod \"5885f5e1-fac5-4980-8501-f882f612fa04\" (UID: \"5885f5e1-fac5-4980-8501-f882f612fa04\") " Dec 27 08:00:03 crc kubenswrapper[4934]: I1227 08:00:03.826065 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5885f5e1-fac5-4980-8501-f882f612fa04-kube-api-access-crhk4" (OuterVolumeSpecName: "kube-api-access-crhk4") pod "5885f5e1-fac5-4980-8501-f882f612fa04" (UID: "5885f5e1-fac5-4980-8501-f882f612fa04"). InnerVolumeSpecName "kube-api-access-crhk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:00:03 crc kubenswrapper[4934]: I1227 08:00:03.826348 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5885f5e1-fac5-4980-8501-f882f612fa04-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5885f5e1-fac5-4980-8501-f882f612fa04" (UID: "5885f5e1-fac5-4980-8501-f882f612fa04"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:00:03 crc kubenswrapper[4934]: I1227 08:00:03.827320 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5885f5e1-fac5-4980-8501-f882f612fa04-config-volume" (OuterVolumeSpecName: "config-volume") pod "5885f5e1-fac5-4980-8501-f882f612fa04" (UID: "5885f5e1-fac5-4980-8501-f882f612fa04"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:00:03 crc kubenswrapper[4934]: I1227 08:00:03.923497 4934 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5885f5e1-fac5-4980-8501-f882f612fa04-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 27 08:00:03 crc kubenswrapper[4934]: I1227 08:00:03.923542 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crhk4\" (UniqueName: \"kubernetes.io/projected/5885f5e1-fac5-4980-8501-f882f612fa04-kube-api-access-crhk4\") on node \"crc\" DevicePath \"\"" Dec 27 08:00:03 crc kubenswrapper[4934]: I1227 08:00:03.923554 4934 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5885f5e1-fac5-4980-8501-f882f612fa04-config-volume\") on node \"crc\" DevicePath \"\"" Dec 27 08:00:04 crc kubenswrapper[4934]: I1227 08:00:04.338569 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg" event={"ID":"5885f5e1-fac5-4980-8501-f882f612fa04","Type":"ContainerDied","Data":"b3a492eb0d415687a6b4ac86ef95499ffabe6225a357bd00b0cd7e9f6cda0496"} Dec 27 08:00:04 crc kubenswrapper[4934]: I1227 08:00:04.338613 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3a492eb0d415687a6b4ac86ef95499ffabe6225a357bd00b0cd7e9f6cda0496" Dec 27 08:00:04 crc kubenswrapper[4934]: I1227 08:00:04.338671 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg" Dec 27 08:00:04 crc kubenswrapper[4934]: I1227 08:00:04.693186 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc"] Dec 27 08:00:04 crc kubenswrapper[4934]: E1227 08:00:04.693566 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4" containerName="pull" Dec 27 08:00:04 crc kubenswrapper[4934]: I1227 08:00:04.693586 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4" containerName="pull" Dec 27 08:00:04 crc kubenswrapper[4934]: E1227 08:00:04.693608 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5885f5e1-fac5-4980-8501-f882f612fa04" containerName="collect-profiles" Dec 27 08:00:04 crc kubenswrapper[4934]: I1227 08:00:04.693617 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="5885f5e1-fac5-4980-8501-f882f612fa04" containerName="collect-profiles" Dec 27 08:00:04 crc kubenswrapper[4934]: E1227 08:00:04.693632 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4" containerName="extract" Dec 27 08:00:04 crc kubenswrapper[4934]: I1227 08:00:04.693640 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4" containerName="extract" Dec 27 08:00:04 crc kubenswrapper[4934]: E1227 08:00:04.693677 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4" containerName="util" Dec 27 08:00:04 crc kubenswrapper[4934]: I1227 08:00:04.693686 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4" containerName="util" Dec 27 08:00:04 crc kubenswrapper[4934]: I1227 08:00:04.693877 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="5885f5e1-fac5-4980-8501-f882f612fa04" containerName="collect-profiles" Dec 27 08:00:04 crc kubenswrapper[4934]: I1227 08:00:04.693903 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4" containerName="extract" Dec 27 08:00:04 crc kubenswrapper[4934]: I1227 08:00:04.694944 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc" Dec 27 08:00:04 crc kubenswrapper[4934]: I1227 08:00:04.697326 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-jv67v" Dec 27 08:00:04 crc kubenswrapper[4934]: I1227 08:00:04.734978 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc"] Dec 27 08:00:04 crc kubenswrapper[4934]: I1227 08:00:04.740625 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v4kj\" (UniqueName: \"kubernetes.io/projected/b461380d-6573-4898-9bc8-c82ceba822d8-kube-api-access-6v4kj\") pod \"openstack-operator-controller-operator-78479c6bd7-j8dhc\" (UID: \"b461380d-6573-4898-9bc8-c82ceba822d8\") " pod="openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc" Dec 27 08:00:04 crc kubenswrapper[4934]: I1227 08:00:04.842756 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v4kj\" (UniqueName: \"kubernetes.io/projected/b461380d-6573-4898-9bc8-c82ceba822d8-kube-api-access-6v4kj\") pod \"openstack-operator-controller-operator-78479c6bd7-j8dhc\" (UID: \"b461380d-6573-4898-9bc8-c82ceba822d8\") " pod="openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc" Dec 27 08:00:04 crc kubenswrapper[4934]: I1227 08:00:04.869702 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v4kj\" (UniqueName: \"kubernetes.io/projected/b461380d-6573-4898-9bc8-c82ceba822d8-kube-api-access-6v4kj\") pod \"openstack-operator-controller-operator-78479c6bd7-j8dhc\" (UID: \"b461380d-6573-4898-9bc8-c82ceba822d8\") " pod="openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc" Dec 27 08:00:05 crc kubenswrapper[4934]: I1227 08:00:05.012668 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc" Dec 27 08:00:05 crc kubenswrapper[4934]: I1227 08:00:05.509641 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc"] Dec 27 08:00:06 crc kubenswrapper[4934]: I1227 08:00:06.361874 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc" event={"ID":"b461380d-6573-4898-9bc8-c82ceba822d8","Type":"ContainerStarted","Data":"0ba6e097fbc69c0730b7ff315323187fa3a129293ea60d1432d9131270013836"} Dec 27 08:00:10 crc kubenswrapper[4934]: I1227 08:00:10.401841 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc" event={"ID":"b461380d-6573-4898-9bc8-c82ceba822d8","Type":"ContainerStarted","Data":"99d594bac3dc404bf5a9ce391065295d53abe364be9cfd3c67e1ef7172872c44"} Dec 27 08:00:13 crc kubenswrapper[4934]: I1227 08:00:13.447691 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc" event={"ID":"b461380d-6573-4898-9bc8-c82ceba822d8","Type":"ContainerStarted","Data":"a0e777ae0aa316be2bb65c0a70e9614e8f1c30c66fb617b2a7ffdb1e083e7afe"} Dec 27 08:00:13 crc kubenswrapper[4934]: I1227 08:00:13.448334 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc" Dec 27 08:00:13 crc kubenswrapper[4934]: I1227 08:00:13.504786 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc" podStartSLOduration=2.66314636 podStartE2EDuration="9.504770415s" podCreationTimestamp="2025-12-27 08:00:04 +0000 UTC" firstStartedPulling="2025-12-27 08:00:05.507200772 +0000 UTC m=+1066.327641376" lastFinishedPulling="2025-12-27 08:00:12.348824837 +0000 UTC m=+1073.169265431" observedRunningTime="2025-12-27 08:00:13.500817435 +0000 UTC m=+1074.321258029" watchObservedRunningTime="2025-12-27 08:00:13.504770415 +0000 UTC m=+1074.325211009" Dec 27 08:00:14 crc kubenswrapper[4934]: I1227 08:00:14.458750 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.261307 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.263183 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.268775 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-9p4nw" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.270053 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.271886 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.273575 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-p8vfc" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.280133 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.281738 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.283860 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-nrscp" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.300172 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.317402 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.327047 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.360584 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-64fb555449-tbmpv"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.362315 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-64fb555449-tbmpv" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.372539 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-rbvfv" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.386867 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmqhx\" (UniqueName: \"kubernetes.io/projected/dc4ddef4-e9a5-4934-846e-1d1bcfe9f223-kube-api-access-dmqhx\") pod \"barbican-operator-controller-manager-649cf68dc9-t4bqz\" (UID: \"dc4ddef4-e9a5-4934-846e-1d1bcfe9f223\") " pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.386940 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjdc5\" (UniqueName: \"kubernetes.io/projected/fc1c7052-d894-4fa1-b6bb-951cad9e32e0-kube-api-access-qjdc5\") pod \"designate-operator-controller-manager-69977bdf55-p5jgx\" (UID: \"fc1c7052-d894-4fa1-b6bb-951cad9e32e0\") " pod="openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.387361 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t7b8\" (UniqueName: \"kubernetes.io/projected/935a4c26-2678-4ead-87ee-1297850d89ff-kube-api-access-6t7b8\") pod \"cinder-operator-controller-manager-669b58f65-t89p4\" (UID: \"935a4c26-2678-4ead-87ee-1297850d89ff\") " pod="openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.392727 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.394159 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.409286 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-j6nvt" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.428383 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-64fb555449-tbmpv"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.450479 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.488861 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjdc5\" (UniqueName: \"kubernetes.io/projected/fc1c7052-d894-4fa1-b6bb-951cad9e32e0-kube-api-access-qjdc5\") pod \"designate-operator-controller-manager-69977bdf55-p5jgx\" (UID: \"fc1c7052-d894-4fa1-b6bb-951cad9e32e0\") " pod="openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.488986 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mggfx\" (UniqueName: \"kubernetes.io/projected/3df60bce-0768-4949-8b8d-3bb28566cd4a-kube-api-access-mggfx\") pod \"glance-operator-controller-manager-64fb555449-tbmpv\" (UID: \"3df60bce-0768-4949-8b8d-3bb28566cd4a\") " pod="openstack-operators/glance-operator-controller-manager-64fb555449-tbmpv" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.489019 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t7b8\" (UniqueName: \"kubernetes.io/projected/935a4c26-2678-4ead-87ee-1297850d89ff-kube-api-access-6t7b8\") pod \"cinder-operator-controller-manager-669b58f65-t89p4\" (UID: \"935a4c26-2678-4ead-87ee-1297850d89ff\") " pod="openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.489107 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pr7j\" (UniqueName: \"kubernetes.io/projected/77029090-da44-403b-a20f-d72105281956-kube-api-access-8pr7j\") pod \"heat-operator-controller-manager-65cdd6cc59-ggc7k\" (UID: \"77029090-da44-403b-a20f-d72105281956\") " pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.489176 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmqhx\" (UniqueName: \"kubernetes.io/projected/dc4ddef4-e9a5-4934-846e-1d1bcfe9f223-kube-api-access-dmqhx\") pod \"barbican-operator-controller-manager-649cf68dc9-t4bqz\" (UID: \"dc4ddef4-e9a5-4934-846e-1d1bcfe9f223\") " pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.515248 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.516731 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.521230 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjdc5\" (UniqueName: \"kubernetes.io/projected/fc1c7052-d894-4fa1-b6bb-951cad9e32e0-kube-api-access-qjdc5\") pod \"designate-operator-controller-manager-69977bdf55-p5jgx\" (UID: \"fc1c7052-d894-4fa1-b6bb-951cad9e32e0\") " pod="openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.524807 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.526193 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.534147 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmqhx\" (UniqueName: \"kubernetes.io/projected/dc4ddef4-e9a5-4934-846e-1d1bcfe9f223-kube-api-access-dmqhx\") pod \"barbican-operator-controller-manager-649cf68dc9-t4bqz\" (UID: \"dc4ddef4-e9a5-4934-846e-1d1bcfe9f223\") " pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.534664 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.534771 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-w9x2m" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.535125 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-gkqb8" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.540581 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t7b8\" (UniqueName: \"kubernetes.io/projected/935a4c26-2678-4ead-87ee-1297850d89ff-kube-api-access-6t7b8\") pod \"cinder-operator-controller-manager-669b58f65-t89p4\" (UID: \"935a4c26-2678-4ead-87ee-1297850d89ff\") " pod="openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.547612 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.570141 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.587919 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.589967 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.591479 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e2db190a-b455-44fc-a43e-2677e5af27b2-cert\") pod \"infra-operator-controller-manager-6c77d5f585-lg74h\" (UID: \"e2db190a-b455-44fc-a43e-2677e5af27b2\") " pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.591528 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mggfx\" (UniqueName: \"kubernetes.io/projected/3df60bce-0768-4949-8b8d-3bb28566cd4a-kube-api-access-mggfx\") pod \"glance-operator-controller-manager-64fb555449-tbmpv\" (UID: \"3df60bce-0768-4949-8b8d-3bb28566cd4a\") " pod="openstack-operators/glance-operator-controller-manager-64fb555449-tbmpv" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.591615 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pr7j\" (UniqueName: \"kubernetes.io/projected/77029090-da44-403b-a20f-d72105281956-kube-api-access-8pr7j\") pod \"heat-operator-controller-manager-65cdd6cc59-ggc7k\" (UID: \"77029090-da44-403b-a20f-d72105281956\") " pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.591837 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpkpp\" (UniqueName: \"kubernetes.io/projected/e2db190a-b455-44fc-a43e-2677e5af27b2-kube-api-access-bpkpp\") pod \"infra-operator-controller-manager-6c77d5f585-lg74h\" (UID: \"e2db190a-b455-44fc-a43e-2677e5af27b2\") " pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.591894 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f92ht\" (UniqueName: \"kubernetes.io/projected/9fbceb4d-e7ae-42c0-8a3f-17807bc9f2ca-kube-api-access-f92ht\") pod \"horizon-operator-controller-manager-54996ddd58-rjkpr\" (UID: \"9fbceb4d-e7ae-42c0-8a3f-17807bc9f2ca\") " pod="openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.593291 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-vz89c" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.600251 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.604189 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.605856 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.608312 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-hcrnx" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.614662 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.619002 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.627209 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pr7j\" (UniqueName: \"kubernetes.io/projected/77029090-da44-403b-a20f-d72105281956-kube-api-access-8pr7j\") pod \"heat-operator-controller-manager-65cdd6cc59-ggc7k\" (UID: \"77029090-da44-403b-a20f-d72105281956\") " pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.640126 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mggfx\" (UniqueName: \"kubernetes.io/projected/3df60bce-0768-4949-8b8d-3bb28566cd4a-kube-api-access-mggfx\") pod \"glance-operator-controller-manager-64fb555449-tbmpv\" (UID: \"3df60bce-0768-4949-8b8d-3bb28566cd4a\") " pod="openstack-operators/glance-operator-controller-manager-64fb555449-tbmpv" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.640579 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.664702 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.670177 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7cc599445b-dlwjl"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.671389 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7cc599445b-dlwjl" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.673694 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-b945r" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.679145 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.680440 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.684166 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7cc599445b-dlwjl"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.684670 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-tl9c8" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.691279 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.694375 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdkct\" (UniqueName: \"kubernetes.io/projected/f41746ba-5695-41bc-8078-29ba3d80e3f1-kube-api-access-rdkct\") pod \"ironic-operator-controller-manager-7564bdd644-bg78w\" (UID: \"f41746ba-5695-41bc-8078-29ba3d80e3f1\") " pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.694431 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpkpp\" (UniqueName: \"kubernetes.io/projected/e2db190a-b455-44fc-a43e-2677e5af27b2-kube-api-access-bpkpp\") pod \"infra-operator-controller-manager-6c77d5f585-lg74h\" (UID: \"e2db190a-b455-44fc-a43e-2677e5af27b2\") " pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.694477 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f92ht\" (UniqueName: \"kubernetes.io/projected/9fbceb4d-e7ae-42c0-8a3f-17807bc9f2ca-kube-api-access-f92ht\") pod \"horizon-operator-controller-manager-54996ddd58-rjkpr\" (UID: \"9fbceb4d-e7ae-42c0-8a3f-17807bc9f2ca\") " pod="openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.694508 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bl9l\" (UniqueName: \"kubernetes.io/projected/718122b6-c9f4-4d6f-92b5-6862c15f4ee6-kube-api-access-4bl9l\") pod \"keystone-operator-controller-manager-7f764db9b-pgjkp\" (UID: \"718122b6-c9f4-4d6f-92b5-6862c15f4ee6\") " pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.694552 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e2db190a-b455-44fc-a43e-2677e5af27b2-cert\") pod \"infra-operator-controller-manager-6c77d5f585-lg74h\" (UID: \"e2db190a-b455-44fc-a43e-2677e5af27b2\") " pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" Dec 27 08:00:52 crc kubenswrapper[4934]: E1227 08:00:52.694785 4934 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 27 08:00:52 crc kubenswrapper[4934]: E1227 08:00:52.694839 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e2db190a-b455-44fc-a43e-2677e5af27b2-cert podName:e2db190a-b455-44fc-a43e-2677e5af27b2 nodeName:}" failed. No retries permitted until 2025-12-27 08:00:53.194822057 +0000 UTC m=+1114.015262651 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e2db190a-b455-44fc-a43e-2677e5af27b2-cert") pod "infra-operator-controller-manager-6c77d5f585-lg74h" (UID: "e2db190a-b455-44fc-a43e-2677e5af27b2") : secret "infra-operator-webhook-server-cert" not found Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.696775 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-64fb555449-tbmpv" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.701830 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-58879495c-d27qh"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.703162 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-58879495c-d27qh" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.714498 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-wlrtd" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.724379 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.726745 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.727432 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpkpp\" (UniqueName: \"kubernetes.io/projected/e2db190a-b455-44fc-a43e-2677e5af27b2-kube-api-access-bpkpp\") pod \"infra-operator-controller-manager-6c77d5f585-lg74h\" (UID: \"e2db190a-b455-44fc-a43e-2677e5af27b2\") " pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.732522 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-8m4j9" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.738295 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-58879495c-d27qh"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.746842 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f92ht\" (UniqueName: \"kubernetes.io/projected/9fbceb4d-e7ae-42c0-8a3f-17807bc9f2ca-kube-api-access-f92ht\") pod \"horizon-operator-controller-manager-54996ddd58-rjkpr\" (UID: \"9fbceb4d-e7ae-42c0-8a3f-17807bc9f2ca\") " pod="openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.753512 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.762712 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.775414 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.777371 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.783663 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-f6vcl" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.798161 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.800068 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bl9l\" (UniqueName: \"kubernetes.io/projected/718122b6-c9f4-4d6f-92b5-6862c15f4ee6-kube-api-access-4bl9l\") pod \"keystone-operator-controller-manager-7f764db9b-pgjkp\" (UID: \"718122b6-c9f4-4d6f-92b5-6862c15f4ee6\") " pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.800150 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpj77\" (UniqueName: \"kubernetes.io/projected/828ebb68-326b-4c32-bc0a-d7f258b45ebd-kube-api-access-fpj77\") pod \"mariadb-operator-controller-manager-64d7c556cd-ffx59\" (UID: \"828ebb68-326b-4c32-bc0a-d7f258b45ebd\") " pod="openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.800228 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfzj2\" (UniqueName: \"kubernetes.io/projected/d090706d-34a7-4c72-9cb9-ad0601db35a6-kube-api-access-dfzj2\") pod \"nova-operator-controller-manager-6b444986fd-djrtj\" (UID: \"d090706d-34a7-4c72-9cb9-ad0601db35a6\") " pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.800277 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxhnw\" (UniqueName: \"kubernetes.io/projected/89ea143d-f68d-46fc-b15a-6e00a418f65c-kube-api-access-jxhnw\") pod \"neutron-operator-controller-manager-58879495c-d27qh\" (UID: \"89ea143d-f68d-46fc-b15a-6e00a418f65c\") " pod="openstack-operators/neutron-operator-controller-manager-58879495c-d27qh" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.800444 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fgz5\" (UniqueName: \"kubernetes.io/projected/37d56061-eab6-43e5-8dec-f59220da5d0d-kube-api-access-5fgz5\") pod \"manila-operator-controller-manager-7cc599445b-dlwjl\" (UID: \"37d56061-eab6-43e5-8dec-f59220da5d0d\") " pod="openstack-operators/manila-operator-controller-manager-7cc599445b-dlwjl" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.800509 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdkct\" (UniqueName: \"kubernetes.io/projected/f41746ba-5695-41bc-8078-29ba3d80e3f1-kube-api-access-rdkct\") pod \"ironic-operator-controller-manager-7564bdd644-bg78w\" (UID: \"f41746ba-5695-41bc-8078-29ba3d80e3f1\") " pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.815318 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.816572 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.825932 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.826256 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-fs5cd" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.831017 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdkct\" (UniqueName: \"kubernetes.io/projected/f41746ba-5695-41bc-8078-29ba3d80e3f1-kube-api-access-rdkct\") pod \"ironic-operator-controller-manager-7564bdd644-bg78w\" (UID: \"f41746ba-5695-41bc-8078-29ba3d80e3f1\") " pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.836778 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bl9l\" (UniqueName: \"kubernetes.io/projected/718122b6-c9f4-4d6f-92b5-6862c15f4ee6-kube-api-access-4bl9l\") pod \"keystone-operator-controller-manager-7f764db9b-pgjkp\" (UID: \"718122b6-c9f4-4d6f-92b5-6862c15f4ee6\") " pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.861023 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.865827 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.868862 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-r655f" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.881259 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.903356 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-cc776f956-cj2c7"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.904489 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b5ea792c-b43f-4221-b7bb-aba3421de3d8-cert\") pod \"openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc\" (UID: \"b5ea792c-b43f-4221-b7bb-aba3421de3d8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.904540 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpj77\" (UniqueName: \"kubernetes.io/projected/828ebb68-326b-4c32-bc0a-d7f258b45ebd-kube-api-access-fpj77\") pod \"mariadb-operator-controller-manager-64d7c556cd-ffx59\" (UID: \"828ebb68-326b-4c32-bc0a-d7f258b45ebd\") " pod="openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.904621 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfzj2\" (UniqueName: \"kubernetes.io/projected/d090706d-34a7-4c72-9cb9-ad0601db35a6-kube-api-access-dfzj2\") pod \"nova-operator-controller-manager-6b444986fd-djrtj\" (UID: \"d090706d-34a7-4c72-9cb9-ad0601db35a6\") " pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.904672 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfd7g\" (UniqueName: \"kubernetes.io/projected/b5ea792c-b43f-4221-b7bb-aba3421de3d8-kube-api-access-rfd7g\") pod \"openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc\" (UID: \"b5ea792c-b43f-4221-b7bb-aba3421de3d8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.904706 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxhnw\" (UniqueName: \"kubernetes.io/projected/89ea143d-f68d-46fc-b15a-6e00a418f65c-kube-api-access-jxhnw\") pod \"neutron-operator-controller-manager-58879495c-d27qh\" (UID: \"89ea143d-f68d-46fc-b15a-6e00a418f65c\") " pod="openstack-operators/neutron-operator-controller-manager-58879495c-d27qh" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.904762 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fgz5\" (UniqueName: \"kubernetes.io/projected/37d56061-eab6-43e5-8dec-f59220da5d0d-kube-api-access-5fgz5\") pod \"manila-operator-controller-manager-7cc599445b-dlwjl\" (UID: \"37d56061-eab6-43e5-8dec-f59220da5d0d\") " pod="openstack-operators/manila-operator-controller-manager-7cc599445b-dlwjl" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.904791 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-cc776f956-cj2c7" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.905306 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdxhk\" (UniqueName: \"kubernetes.io/projected/1c0960f0-ef55-479a-b375-a09e69254743-kube-api-access-tdxhk\") pod \"octavia-operator-controller-manager-d5fb87cb8-w2xgr\" (UID: \"1c0960f0-ef55-479a-b375-a09e69254743\") " pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.910558 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-95ngv" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.911241 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.922736 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.929154 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpj77\" (UniqueName: \"kubernetes.io/projected/828ebb68-326b-4c32-bc0a-d7f258b45ebd-kube-api-access-fpj77\") pod \"mariadb-operator-controller-manager-64d7c556cd-ffx59\" (UID: \"828ebb68-326b-4c32-bc0a-d7f258b45ebd\") " pod="openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.933546 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.935474 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.935968 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fgz5\" (UniqueName: \"kubernetes.io/projected/37d56061-eab6-43e5-8dec-f59220da5d0d-kube-api-access-5fgz5\") pod \"manila-operator-controller-manager-7cc599445b-dlwjl\" (UID: \"37d56061-eab6-43e5-8dec-f59220da5d0d\") " pod="openstack-operators/manila-operator-controller-manager-7cc599445b-dlwjl" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.936165 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxhnw\" (UniqueName: \"kubernetes.io/projected/89ea143d-f68d-46fc-b15a-6e00a418f65c-kube-api-access-jxhnw\") pod \"neutron-operator-controller-manager-58879495c-d27qh\" (UID: \"89ea143d-f68d-46fc-b15a-6e00a418f65c\") " pod="openstack-operators/neutron-operator-controller-manager-58879495c-d27qh" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.936614 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfzj2\" (UniqueName: \"kubernetes.io/projected/d090706d-34a7-4c72-9cb9-ad0601db35a6-kube-api-access-dfzj2\") pod \"nova-operator-controller-manager-6b444986fd-djrtj\" (UID: \"d090706d-34a7-4c72-9cb9-ad0601db35a6\") " pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.942202 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg"] Dec 27 08:00:52 crc kubenswrapper[4934]: I1227 08:00:52.942311 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-d6qgt" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.019310 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-cc776f956-cj2c7"] Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.027920 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdxhk\" (UniqueName: \"kubernetes.io/projected/1c0960f0-ef55-479a-b375-a09e69254743-kube-api-access-tdxhk\") pod \"octavia-operator-controller-manager-d5fb87cb8-w2xgr\" (UID: \"1c0960f0-ef55-479a-b375-a09e69254743\") " pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.028138 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwnzv\" (UniqueName: \"kubernetes.io/projected/4f825c5b-90ef-4b56-a759-d7ccbb312bed-kube-api-access-pwnzv\") pod \"swift-operator-controller-manager-7c9ff8845d-ghvpg\" (UID: \"4f825c5b-90ef-4b56-a759-d7ccbb312bed\") " pod="openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.028184 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b5ea792c-b43f-4221-b7bb-aba3421de3d8-cert\") pod \"openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc\" (UID: \"b5ea792c-b43f-4221-b7bb-aba3421de3d8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.028242 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxh8n\" (UniqueName: \"kubernetes.io/projected/59d65d09-c15c-49d9-8ee7-472ffe0dda55-kube-api-access-lxh8n\") pod \"ovn-operator-controller-manager-5b67cfc8fb-6r7n4\" (UID: \"59d65d09-c15c-49d9-8ee7-472ffe0dda55\") " pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.028360 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfd7g\" (UniqueName: \"kubernetes.io/projected/b5ea792c-b43f-4221-b7bb-aba3421de3d8-kube-api-access-rfd7g\") pod \"openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc\" (UID: \"b5ea792c-b43f-4221-b7bb-aba3421de3d8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.028432 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75ff2\" (UniqueName: \"kubernetes.io/projected/7d49c3c9-83c2-4c42-a318-c9d41ddf1bf8-kube-api-access-75ff2\") pod \"placement-operator-controller-manager-cc776f956-cj2c7\" (UID: \"7d49c3c9-83c2-4c42-a318-c9d41ddf1bf8\") " pod="openstack-operators/placement-operator-controller-manager-cc776f956-cj2c7" Dec 27 08:00:53 crc kubenswrapper[4934]: E1227 08:00:53.029238 4934 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 27 08:00:53 crc kubenswrapper[4934]: E1227 08:00:53.029313 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5ea792c-b43f-4221-b7bb-aba3421de3d8-cert podName:b5ea792c-b43f-4221-b7bb-aba3421de3d8 nodeName:}" failed. No retries permitted until 2025-12-27 08:00:53.529283406 +0000 UTC m=+1114.349724000 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b5ea792c-b43f-4221-b7bb-aba3421de3d8-cert") pod "openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" (UID: "b5ea792c-b43f-4221-b7bb-aba3421de3d8") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.063772 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdxhk\" (UniqueName: \"kubernetes.io/projected/1c0960f0-ef55-479a-b375-a09e69254743-kube-api-access-tdxhk\") pod \"octavia-operator-controller-manager-d5fb87cb8-w2xgr\" (UID: \"1c0960f0-ef55-479a-b375-a09e69254743\") " pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.088438 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.090942 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.092483 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfd7g\" (UniqueName: \"kubernetes.io/projected/b5ea792c-b43f-4221-b7bb-aba3421de3d8-kube-api-access-rfd7g\") pod \"openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc\" (UID: \"b5ea792c-b43f-4221-b7bb-aba3421de3d8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.105978 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.106821 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7cc599445b-dlwjl" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.147401 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75ff2\" (UniqueName: \"kubernetes.io/projected/7d49c3c9-83c2-4c42-a318-c9d41ddf1bf8-kube-api-access-75ff2\") pod \"placement-operator-controller-manager-cc776f956-cj2c7\" (UID: \"7d49c3c9-83c2-4c42-a318-c9d41ddf1bf8\") " pod="openstack-operators/placement-operator-controller-manager-cc776f956-cj2c7" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.149908 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwnzv\" (UniqueName: \"kubernetes.io/projected/4f825c5b-90ef-4b56-a759-d7ccbb312bed-kube-api-access-pwnzv\") pod \"swift-operator-controller-manager-7c9ff8845d-ghvpg\" (UID: \"4f825c5b-90ef-4b56-a759-d7ccbb312bed\") " pod="openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.150004 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxh8n\" (UniqueName: \"kubernetes.io/projected/59d65d09-c15c-49d9-8ee7-472ffe0dda55-kube-api-access-lxh8n\") pod \"ovn-operator-controller-manager-5b67cfc8fb-6r7n4\" (UID: \"59d65d09-c15c-49d9-8ee7-472ffe0dda55\") " pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.155359 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj"] Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.169073 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxh8n\" (UniqueName: \"kubernetes.io/projected/59d65d09-c15c-49d9-8ee7-472ffe0dda55-kube-api-access-lxh8n\") pod \"ovn-operator-controller-manager-5b67cfc8fb-6r7n4\" (UID: \"59d65d09-c15c-49d9-8ee7-472ffe0dda55\") " pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.169317 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj"] Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.170257 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.172780 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwnzv\" (UniqueName: \"kubernetes.io/projected/4f825c5b-90ef-4b56-a759-d7ccbb312bed-kube-api-access-pwnzv\") pod \"swift-operator-controller-manager-7c9ff8845d-ghvpg\" (UID: \"4f825c5b-90ef-4b56-a759-d7ccbb312bed\") " pod="openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.176403 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-58879495c-d27qh" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.194901 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-kfnvq" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.198140 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75ff2\" (UniqueName: \"kubernetes.io/projected/7d49c3c9-83c2-4c42-a318-c9d41ddf1bf8-kube-api-access-75ff2\") pod \"placement-operator-controller-manager-cc776f956-cj2c7\" (UID: \"7d49c3c9-83c2-4c42-a318-c9d41ddf1bf8\") " pod="openstack-operators/placement-operator-controller-manager-cc776f956-cj2c7" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.227455 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.236160 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.246133 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm"] Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.247628 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.252976 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-bgb42" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.253945 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e2db190a-b455-44fc-a43e-2677e5af27b2-cert\") pod \"infra-operator-controller-manager-6c77d5f585-lg74h\" (UID: \"e2db190a-b455-44fc-a43e-2677e5af27b2\") " pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.254000 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjf9r\" (UniqueName: \"kubernetes.io/projected/f2d5e345-1bd9-4313-bdae-7637cb7ed944-kube-api-access-qjf9r\") pod \"telemetry-operator-controller-manager-6659cb75fc-9kzpj\" (UID: \"f2d5e345-1bd9-4313-bdae-7637cb7ed944\") " pod="openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.256938 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm"] Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.259464 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e2db190a-b455-44fc-a43e-2677e5af27b2-cert\") pod \"infra-operator-controller-manager-6c77d5f585-lg74h\" (UID: \"e2db190a-b455-44fc-a43e-2677e5af27b2\") " pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.309735 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.331238 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-cc776f956-cj2c7" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.331774 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.356026 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjf9r\" (UniqueName: \"kubernetes.io/projected/f2d5e345-1bd9-4313-bdae-7637cb7ed944-kube-api-access-qjf9r\") pod \"telemetry-operator-controller-manager-6659cb75fc-9kzpj\" (UID: \"f2d5e345-1bd9-4313-bdae-7637cb7ed944\") " pod="openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.356971 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5wt4\" (UniqueName: \"kubernetes.io/projected/f1cce183-e0a8-4cf1-854e-7a60986f194f-kube-api-access-b5wt4\") pod \"test-operator-controller-manager-5d79c6465c-zldvm\" (UID: \"f1cce183-e0a8-4cf1-854e-7a60986f194f\") " pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.362653 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln"] Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.364421 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.367532 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-cv8ml" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.380404 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjf9r\" (UniqueName: \"kubernetes.io/projected/f2d5e345-1bd9-4313-bdae-7637cb7ed944-kube-api-access-qjf9r\") pod \"telemetry-operator-controller-manager-6659cb75fc-9kzpj\" (UID: \"f2d5e345-1bd9-4313-bdae-7637cb7ed944\") " pod="openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.386244 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln"] Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.392668 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.459400 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5wt4\" (UniqueName: \"kubernetes.io/projected/f1cce183-e0a8-4cf1-854e-7a60986f194f-kube-api-access-b5wt4\") pod \"test-operator-controller-manager-5d79c6465c-zldvm\" (UID: \"f1cce183-e0a8-4cf1-854e-7a60986f194f\") " pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.459847 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wvrl\" (UniqueName: \"kubernetes.io/projected/35d6eb8d-3797-432c-a5ea-65f31adf7262-kube-api-access-4wvrl\") pod \"watcher-operator-controller-manager-59cf64b5bd-28lln\" (UID: \"35d6eb8d-3797-432c-a5ea-65f31adf7262\") " pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.460019 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf"] Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.461466 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.463633 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.463842 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-fpj7r" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.485124 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5wt4\" (UniqueName: \"kubernetes.io/projected/f1cce183-e0a8-4cf1-854e-7a60986f194f-kube-api-access-b5wt4\") pod \"test-operator-controller-manager-5d79c6465c-zldvm\" (UID: \"f1cce183-e0a8-4cf1-854e-7a60986f194f\") " pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.493931 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf"] Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.499738 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z9x75"] Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.502033 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z9x75" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.505225 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-2qqgq" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.508103 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z9x75"] Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.525444 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.525555 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4"] Dec 27 08:00:53 crc kubenswrapper[4934]: W1227 08:00:53.557755 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc4ddef4_e9a5_4934_846e_1d1bcfe9f223.slice/crio-c17f44c91caf0ddd2d5736788f547892a0c1e4a0ade763118becf9389520ae30 WatchSource:0}: Error finding container c17f44c91caf0ddd2d5736788f547892a0c1e4a0ade763118becf9389520ae30: Status 404 returned error can't find the container with id c17f44c91caf0ddd2d5736788f547892a0c1e4a0ade763118becf9389520ae30 Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.561657 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e44e23aa-f466-41cb-b855-d0bdc25cf05b-cert\") pod \"openstack-operator-controller-manager-5758858f4d-mg9wf\" (UID: \"e44e23aa-f466-41cb-b855-d0bdc25cf05b\") " pod="openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.561805 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b5ea792c-b43f-4221-b7bb-aba3421de3d8-cert\") pod \"openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc\" (UID: \"b5ea792c-b43f-4221-b7bb-aba3421de3d8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.561832 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwf46\" (UniqueName: \"kubernetes.io/projected/e44e23aa-f466-41cb-b855-d0bdc25cf05b-kube-api-access-fwf46\") pod \"openstack-operator-controller-manager-5758858f4d-mg9wf\" (UID: \"e44e23aa-f466-41cb-b855-d0bdc25cf05b\") " pod="openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.561855 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wvrl\" (UniqueName: \"kubernetes.io/projected/35d6eb8d-3797-432c-a5ea-65f31adf7262-kube-api-access-4wvrl\") pod \"watcher-operator-controller-manager-59cf64b5bd-28lln\" (UID: \"35d6eb8d-3797-432c-a5ea-65f31adf7262\") " pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.575505 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.587992 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b5ea792c-b43f-4221-b7bb-aba3421de3d8-cert\") pod \"openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc\" (UID: \"b5ea792c-b43f-4221-b7bb-aba3421de3d8\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.591330 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.595810 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wvrl\" (UniqueName: \"kubernetes.io/projected/35d6eb8d-3797-432c-a5ea-65f31adf7262-kube-api-access-4wvrl\") pod \"watcher-operator-controller-manager-59cf64b5bd-28lln\" (UID: \"35d6eb8d-3797-432c-a5ea-65f31adf7262\") " pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.605245 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-64fb555449-tbmpv"] Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.625712 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz"] Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.663119 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77m2t\" (UniqueName: \"kubernetes.io/projected/e5e99871-c4b8-4bde-b7b7-41b7b5e2f5b0-kube-api-access-77m2t\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-z9x75\" (UID: \"e5e99871-c4b8-4bde-b7b7-41b7b5e2f5b0\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z9x75" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.663180 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwf46\" (UniqueName: \"kubernetes.io/projected/e44e23aa-f466-41cb-b855-d0bdc25cf05b-kube-api-access-fwf46\") pod \"openstack-operator-controller-manager-5758858f4d-mg9wf\" (UID: \"e44e23aa-f466-41cb-b855-d0bdc25cf05b\") " pod="openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.663261 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e44e23aa-f466-41cb-b855-d0bdc25cf05b-cert\") pod \"openstack-operator-controller-manager-5758858f4d-mg9wf\" (UID: \"e44e23aa-f466-41cb-b855-d0bdc25cf05b\") " pod="openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.668658 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e44e23aa-f466-41cb-b855-d0bdc25cf05b-cert\") pod \"openstack-operator-controller-manager-5758858f4d-mg9wf\" (UID: \"e44e23aa-f466-41cb-b855-d0bdc25cf05b\") " pod="openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.682587 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwf46\" (UniqueName: \"kubernetes.io/projected/e44e23aa-f466-41cb-b855-d0bdc25cf05b-kube-api-access-fwf46\") pod \"openstack-operator-controller-manager-5758858f4d-mg9wf\" (UID: \"e44e23aa-f466-41cb-b855-d0bdc25cf05b\") " pod="openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.697830 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.733720 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx"] Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.743176 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k"] Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.768251 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr"] Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.770791 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77m2t\" (UniqueName: \"kubernetes.io/projected/e5e99871-c4b8-4bde-b7b7-41b7b5e2f5b0-kube-api-access-77m2t\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-z9x75\" (UID: \"e5e99871-c4b8-4bde-b7b7-41b7b5e2f5b0\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z9x75" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.784600 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.801799 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77m2t\" (UniqueName: \"kubernetes.io/projected/e5e99871-c4b8-4bde-b7b7-41b7b5e2f5b0-kube-api-access-77m2t\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-z9x75\" (UID: \"e5e99871-c4b8-4bde-b7b7-41b7b5e2f5b0\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z9x75" Dec 27 08:00:53 crc kubenswrapper[4934]: W1227 08:00:53.806113 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod77029090_da44_403b_a20f_d72105281956.slice/crio-e2ef8bbf835b6d26e5ac35945d9e5961c385bec0a726b2e05b406fb751adc21c WatchSource:0}: Error finding container e2ef8bbf835b6d26e5ac35945d9e5961c385bec0a726b2e05b406fb751adc21c: Status 404 returned error can't find the container with id e2ef8bbf835b6d26e5ac35945d9e5961c385bec0a726b2e05b406fb751adc21c Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.826876 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z9x75" Dec 27 08:00:53 crc kubenswrapper[4934]: I1227 08:00:53.919646 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp"] Dec 27 08:00:54 crc kubenswrapper[4934]: I1227 08:00:54.294921 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59"] Dec 27 08:00:54 crc kubenswrapper[4934]: I1227 08:00:54.396836 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w"] Dec 27 08:00:54 crc kubenswrapper[4934]: I1227 08:00:54.401689 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr" event={"ID":"9fbceb4d-e7ae-42c0-8a3f-17807bc9f2ca","Type":"ContainerStarted","Data":"d68c5880bc61263c0407879ff13b746a55547a23ecd62f7eaee40e6dc6dd0507"} Dec 27 08:00:54 crc kubenswrapper[4934]: I1227 08:00:54.406570 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-64fb555449-tbmpv" event={"ID":"3df60bce-0768-4949-8b8d-3bb28566cd4a","Type":"ContainerStarted","Data":"ddfdb6177aa4b28f211683d1c87a4eca9d0a32e1265330456847f4b60116f1ce"} Dec 27 08:00:54 crc kubenswrapper[4934]: I1227 08:00:54.409530 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" event={"ID":"77029090-da44-403b-a20f-d72105281956","Type":"ContainerStarted","Data":"e2ef8bbf835b6d26e5ac35945d9e5961c385bec0a726b2e05b406fb751adc21c"} Dec 27 08:00:54 crc kubenswrapper[4934]: I1227 08:00:54.411791 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4" event={"ID":"935a4c26-2678-4ead-87ee-1297850d89ff","Type":"ContainerStarted","Data":"bbae0739c8a92ccb49f0774ecb50912d13bf9ba18448e6b27ec5295fdbb7bc27"} Dec 27 08:00:54 crc kubenswrapper[4934]: I1227 08:00:54.412769 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" event={"ID":"718122b6-c9f4-4d6f-92b5-6862c15f4ee6","Type":"ContainerStarted","Data":"7d7eb0816521f6aba05e6e33c75f0806c1989b372615193d86a4b11f00612fbb"} Dec 27 08:00:54 crc kubenswrapper[4934]: I1227 08:00:54.416532 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59" event={"ID":"828ebb68-326b-4c32-bc0a-d7f258b45ebd","Type":"ContainerStarted","Data":"7ff7bba0844d4b8943255461e5ffbd1f849ba929682a42395ca3810ab5aea877"} Dec 27 08:00:54 crc kubenswrapper[4934]: I1227 08:00:54.427517 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx" event={"ID":"fc1c7052-d894-4fa1-b6bb-951cad9e32e0","Type":"ContainerStarted","Data":"f904eebcfe5d21cf6fdcce0ed17024d0f1c4cebb414f7906081b948b75179575"} Dec 27 08:00:54 crc kubenswrapper[4934]: I1227 08:00:54.429734 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" event={"ID":"dc4ddef4-e9a5-4934-846e-1d1bcfe9f223","Type":"ContainerStarted","Data":"c17f44c91caf0ddd2d5736788f547892a0c1e4a0ade763118becf9389520ae30"} Dec 27 08:00:54 crc kubenswrapper[4934]: W1227 08:00:54.446858 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37d56061_eab6_43e5_8dec_f59220da5d0d.slice/crio-29e6107f1c362e3d0bf11342d1fbc87ae9367996a980e82bfcd679830b9deb5e WatchSource:0}: Error finding container 29e6107f1c362e3d0bf11342d1fbc87ae9367996a980e82bfcd679830b9deb5e: Status 404 returned error can't find the container with id 29e6107f1c362e3d0bf11342d1fbc87ae9367996a980e82bfcd679830b9deb5e Dec 27 08:00:54 crc kubenswrapper[4934]: I1227 08:00:54.449766 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-58879495c-d27qh"] Dec 27 08:00:54 crc kubenswrapper[4934]: I1227 08:00:54.466725 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7cc599445b-dlwjl"] Dec 27 08:00:54 crc kubenswrapper[4934]: I1227 08:00:54.475503 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj"] Dec 27 08:00:54 crc kubenswrapper[4934]: W1227 08:00:54.485018 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59d65d09_c15c_49d9_8ee7_472ffe0dda55.slice/crio-76c41cb15d01b8096fb46b6e84789322b86c34a61ac5f5706193637ffbaab56a WatchSource:0}: Error finding container 76c41cb15d01b8096fb46b6e84789322b86c34a61ac5f5706193637ffbaab56a: Status 404 returned error can't find the container with id 76c41cb15d01b8096fb46b6e84789322b86c34a61ac5f5706193637ffbaab56a Dec 27 08:00:54 crc kubenswrapper[4934]: I1227 08:00:54.485075 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr"] Dec 27 08:00:54 crc kubenswrapper[4934]: I1227 08:00:54.487841 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4"] Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.099716 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj"] Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.115397 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h"] Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.132619 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-cc776f956-cj2c7"] Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.135270 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg"] Dec 27 08:00:55 crc kubenswrapper[4934]: W1227 08:00:55.157036 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2db190a_b455_44fc_a43e_2677e5af27b2.slice/crio-df11a19b6e67fb82a968a2a3a8f7473f02c00c42808045a6a53d3aacc0aea550 WatchSource:0}: Error finding container df11a19b6e67fb82a968a2a3a8f7473f02c00c42808045a6a53d3aacc0aea550: Status 404 returned error can't find the container with id df11a19b6e67fb82a968a2a3a8f7473f02c00c42808045a6a53d3aacc0aea550 Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.166110 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z9x75"] Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.175969 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln"] Dec 27 08:00:55 crc kubenswrapper[4934]: E1227 08:00:55.179758 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4bb51fe2697380ac96838a2d69e937ab8308587169eed75ed5922636156fb27e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4wvrl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-59cf64b5bd-28lln_openstack-operators(35d6eb8d-3797-432c-a5ea-65f31adf7262): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.185733 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf"] Dec 27 08:00:55 crc kubenswrapper[4934]: W1227 08:00:55.192867 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5ea792c_b43f_4221_b7bb_aba3421de3d8.slice/crio-fd14f8ece24bbc44f2290c793e432c24dc2e1204c1b98346cde119cd2bb94f35 WatchSource:0}: Error finding container fd14f8ece24bbc44f2290c793e432c24dc2e1204c1b98346cde119cd2bb94f35: Status 404 returned error can't find the container with id fd14f8ece24bbc44f2290c793e432c24dc2e1204c1b98346cde119cd2bb94f35 Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.194151 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm"] Dec 27 08:00:55 crc kubenswrapper[4934]: E1227 08:00:55.197306 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bfb1e0635f87094bee949f00fea37cbc27b88c42a7cef1909e0b68e5abd185c7,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:18.0-fr4-latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:18.0-fr4-latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:18.0-fr4-latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rfd7g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc_openstack-operators(b5ea792c-b43f-4221-b7bb-aba3421de3d8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.204363 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc"] Dec 27 08:00:55 crc kubenswrapper[4934]: W1227 08:00:55.215543 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1cce183_e0a8_4cf1_854e_7a60986f194f.slice/crio-c36f09ba7a7527d183809c4fc88f16ea3c7ec2e5ba01b530a61f5509a2b03294 WatchSource:0}: Error finding container c36f09ba7a7527d183809c4fc88f16ea3c7ec2e5ba01b530a61f5509a2b03294: Status 404 returned error can't find the container with id c36f09ba7a7527d183809c4fc88f16ea3c7ec2e5ba01b530a61f5509a2b03294 Dec 27 08:00:55 crc kubenswrapper[4934]: E1227 08:00:55.219928 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:b9e09dbcf7f70960e90ecbb8b31bbb7acf141fc4975f69e37482df2bd0ea2773,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b5wt4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5d79c6465c-zldvm_openstack-operators(f1cce183-e0a8-4cf1-854e-7a60986f194f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 27 08:00:55 crc kubenswrapper[4934]: E1227 08:00:55.438448 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" podUID="35d6eb8d-3797-432c-a5ea-65f31adf7262" Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.516651 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cc599445b-dlwjl" event={"ID":"37d56061-eab6-43e5-8dec-f59220da5d0d","Type":"ContainerStarted","Data":"29e6107f1c362e3d0bf11342d1fbc87ae9367996a980e82bfcd679830b9deb5e"} Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.541273 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" event={"ID":"1c0960f0-ef55-479a-b375-a09e69254743","Type":"ContainerStarted","Data":"559ba8be2742d75979d59e612eaa7b4f703b26a5161d4b7258a3c8831ba3db2d"} Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.546258 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf" event={"ID":"e44e23aa-f466-41cb-b855-d0bdc25cf05b","Type":"ContainerStarted","Data":"fc1efca29f28c419ae959cbb6849958dec2facd1e61ecec7a5cb8f436ad1b511"} Dec 27 08:00:55 crc kubenswrapper[4934]: E1227 08:00:55.598071 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" podUID="b5ea792c-b43f-4221-b7bb-aba3421de3d8" Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.617320 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" event={"ID":"f41746ba-5695-41bc-8078-29ba3d80e3f1","Type":"ContainerStarted","Data":"9ca4d8222caef1ac47a53a3cb0ada9b2bb25580f7246fc2f88de0a1b428f2790"} Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.687033 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-58879495c-d27qh" event={"ID":"89ea143d-f68d-46fc-b15a-6e00a418f65c","Type":"ContainerStarted","Data":"d78bc53b9a30c8d758ce3d46636da566bab23d34508e855c39c61f8244999cbd"} Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.726509 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" event={"ID":"e2db190a-b455-44fc-a43e-2677e5af27b2","Type":"ContainerStarted","Data":"df11a19b6e67fb82a968a2a3a8f7473f02c00c42808045a6a53d3aacc0aea550"} Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.759402 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" event={"ID":"b5ea792c-b43f-4221-b7bb-aba3421de3d8","Type":"ContainerStarted","Data":"fd14f8ece24bbc44f2290c793e432c24dc2e1204c1b98346cde119cd2bb94f35"} Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.770416 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj" event={"ID":"f2d5e345-1bd9-4313-bdae-7637cb7ed944","Type":"ContainerStarted","Data":"f061bead34f352ab4dd81626a3c622cfd434f7086e039fdb757de16752bda234"} Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.782782 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" event={"ID":"59d65d09-c15c-49d9-8ee7-472ffe0dda55","Type":"ContainerStarted","Data":"76c41cb15d01b8096fb46b6e84789322b86c34a61ac5f5706193637ffbaab56a"} Dec 27 08:00:55 crc kubenswrapper[4934]: E1227 08:00:55.794703 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bfb1e0635f87094bee949f00fea37cbc27b88c42a7cef1909e0b68e5abd185c7\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" podUID="b5ea792c-b43f-4221-b7bb-aba3421de3d8" Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.812044 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" event={"ID":"d090706d-34a7-4c72-9cb9-ad0601db35a6","Type":"ContainerStarted","Data":"7718cba7d87f7e21c5dfcf9a549629597fd6da548fdc4e4501ccc86e6f01c4ed"} Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.815252 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z9x75" event={"ID":"e5e99871-c4b8-4bde-b7b7-41b7b5e2f5b0","Type":"ContainerStarted","Data":"1283dd412cbe997001b6ccb77fb931f0492f917063fa803be81ec99c8346a40d"} Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.816824 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-cc776f956-cj2c7" event={"ID":"7d49c3c9-83c2-4c42-a318-c9d41ddf1bf8","Type":"ContainerStarted","Data":"44516046e6cd0efb5fa9224c827871885f08cda3931917321a79b4b5ea6cc2e1"} Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.818511 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" event={"ID":"35d6eb8d-3797-432c-a5ea-65f31adf7262","Type":"ContainerStarted","Data":"e8a8adf4c60f2a36f89b08f08b93088e4bb0fd85f4df77ec085babcfc5fd5f6a"} Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.818544 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" event={"ID":"35d6eb8d-3797-432c-a5ea-65f31adf7262","Type":"ContainerStarted","Data":"425e27c63b84c0bf792cf3ce83b6e8eac1f9e8b1e8afe053c8deaacad282d3b3"} Dec 27 08:00:55 crc kubenswrapper[4934]: E1227 08:00:55.821110 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4bb51fe2697380ac96838a2d69e937ab8308587169eed75ed5922636156fb27e\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" podUID="35d6eb8d-3797-432c-a5ea-65f31adf7262" Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.821983 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" event={"ID":"f1cce183-e0a8-4cf1-854e-7a60986f194f","Type":"ContainerStarted","Data":"c36f09ba7a7527d183809c4fc88f16ea3c7ec2e5ba01b530a61f5509a2b03294"} Dec 27 08:00:55 crc kubenswrapper[4934]: I1227 08:00:55.883610 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg" event={"ID":"4f825c5b-90ef-4b56-a759-d7ccbb312bed","Type":"ContainerStarted","Data":"e7c360c85c7e7081b0f57a88656d37daa5f9d6f615ba939db548f818c8e2d13d"} Dec 27 08:00:55 crc kubenswrapper[4934]: E1227 08:00:55.889239 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" podUID="f1cce183-e0a8-4cf1-854e-7a60986f194f" Dec 27 08:00:56 crc kubenswrapper[4934]: I1227 08:00:56.908683 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf" event={"ID":"e44e23aa-f466-41cb-b855-d0bdc25cf05b","Type":"ContainerStarted","Data":"8e50bd381ed20343a8a4c3f912567677789ee40b137ae904c94c43e16927fc80"} Dec 27 08:00:56 crc kubenswrapper[4934]: I1227 08:00:56.908725 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf" event={"ID":"e44e23aa-f466-41cb-b855-d0bdc25cf05b","Type":"ContainerStarted","Data":"5204c773f8058dd2ccc7ae35f4ff843081bfb10eaa682a389c50352b019089b1"} Dec 27 08:00:56 crc kubenswrapper[4934]: I1227 08:00:56.909692 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf" Dec 27 08:00:56 crc kubenswrapper[4934]: I1227 08:00:56.917053 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" event={"ID":"b5ea792c-b43f-4221-b7bb-aba3421de3d8","Type":"ContainerStarted","Data":"91be2205b2555697ece1fd56fffeb54dbfdbedf0a9f4bf5c75d6f4ca53086431"} Dec 27 08:00:56 crc kubenswrapper[4934]: E1227 08:00:56.918950 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bfb1e0635f87094bee949f00fea37cbc27b88c42a7cef1909e0b68e5abd185c7\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" podUID="b5ea792c-b43f-4221-b7bb-aba3421de3d8" Dec 27 08:00:56 crc kubenswrapper[4934]: I1227 08:00:56.921321 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" event={"ID":"f1cce183-e0a8-4cf1-854e-7a60986f194f","Type":"ContainerStarted","Data":"802da22a0f496da23dd03910eb5bdb348f693c7a5318243f1507db22404add71"} Dec 27 08:00:56 crc kubenswrapper[4934]: E1227 08:00:56.925042 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:b9e09dbcf7f70960e90ecbb8b31bbb7acf141fc4975f69e37482df2bd0ea2773\\\"\"" pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" podUID="f1cce183-e0a8-4cf1-854e-7a60986f194f" Dec 27 08:00:56 crc kubenswrapper[4934]: E1227 08:00:56.925121 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4bb51fe2697380ac96838a2d69e937ab8308587169eed75ed5922636156fb27e\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" podUID="35d6eb8d-3797-432c-a5ea-65f31adf7262" Dec 27 08:00:56 crc kubenswrapper[4934]: I1227 08:00:56.938105 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf" podStartSLOduration=3.938089392 podStartE2EDuration="3.938089392s" podCreationTimestamp="2025-12-27 08:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:00:56.932607104 +0000 UTC m=+1117.753047698" watchObservedRunningTime="2025-12-27 08:00:56.938089392 +0000 UTC m=+1117.758529986" Dec 27 08:00:57 crc kubenswrapper[4934]: E1227 08:00:57.933613 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:bfb1e0635f87094bee949f00fea37cbc27b88c42a7cef1909e0b68e5abd185c7\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" podUID="b5ea792c-b43f-4221-b7bb-aba3421de3d8" Dec 27 08:00:57 crc kubenswrapper[4934]: E1227 08:00:57.934811 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:b9e09dbcf7f70960e90ecbb8b31bbb7acf141fc4975f69e37482df2bd0ea2773\\\"\"" pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" podUID="f1cce183-e0a8-4cf1-854e-7a60986f194f" Dec 27 08:01:03 crc kubenswrapper[4934]: I1227 08:01:03.794889 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf" Dec 27 08:01:12 crc kubenswrapper[4934]: E1227 08:01:12.361879 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:2c837009de6475bc22534827c03df6d8649277b71f1c30de2087b6c52aafb326" Dec 27 08:01:12 crc kubenswrapper[4934]: E1227 08:01:12.362484 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:2c837009de6475bc22534827c03df6d8649277b71f1c30de2087b6c52aafb326,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lxh8n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-5b67cfc8fb-6r7n4_openstack-operators(59d65d09-c15c-49d9-8ee7-472ffe0dda55): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:01:12 crc kubenswrapper[4934]: E1227 08:01:12.760027 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:036365e19c95dcd1bb32b3cd45eed8048b454c1f74798bb9df6fb421d2bc0674" Dec 27 08:01:12 crc kubenswrapper[4934]: E1227 08:01:12.760287 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:036365e19c95dcd1bb32b3cd45eed8048b454c1f74798bb9df6fb421d2bc0674,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bpkpp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-6c77d5f585-lg74h_openstack-operators(e2db190a-b455-44fc-a43e-2677e5af27b2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:01:13 crc kubenswrapper[4934]: E1227 08:01:13.173778 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:21ae046934b4ea85d21bdcc1d6b5bc7cb393e319b6dc7bea7ac1cc96aa4a599d" Dec 27 08:01:13 crc kubenswrapper[4934]: E1227 08:01:13.174224 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:21ae046934b4ea85d21bdcc1d6b5bc7cb393e319b6dc7bea7ac1cc96aa4a599d,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pwnzv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-7c9ff8845d-ghvpg_openstack-operators(4f825c5b-90ef-4b56-a759-d7ccbb312bed): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:01:15 crc kubenswrapper[4934]: I1227 08:01:15.330358 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:01:15 crc kubenswrapper[4934]: I1227 08:01:15.330713 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:01:15 crc kubenswrapper[4934]: E1227 08:01:15.392476 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:317bf4faf2283cfc56db7596816bac63f7a6ecd5c7a83eec2ba0ea2f2b8c5448" Dec 27 08:01:15 crc kubenswrapper[4934]: E1227 08:01:15.392755 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:317bf4faf2283cfc56db7596816bac63f7a6ecd5c7a83eec2ba0ea2f2b8c5448,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rdkct,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-7564bdd644-bg78w_openstack-operators(f41746ba-5695-41bc-8078-29ba3d80e3f1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:01:15 crc kubenswrapper[4934]: E1227 08:01:15.833455 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:2c4fe20e044dd8ea1f60f2f3f5e3844d932b4b79439835bd8771c73f16b38312" Dec 27 08:01:15 crc kubenswrapper[4934]: E1227 08:01:15.834015 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:2c4fe20e044dd8ea1f60f2f3f5e3844d932b4b79439835bd8771c73f16b38312,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fpj77,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-64d7c556cd-ffx59_openstack-operators(828ebb68-326b-4c32-bc0a-d7f258b45ebd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:01:17 crc kubenswrapper[4934]: E1227 08:01:17.584470 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:5b7d3d7e5bdd0f7c2ad742990da2488e576cf0ea6dee2e3245192a89cc959096" Dec 27 08:01:17 crc kubenswrapper[4934]: E1227 08:01:17.584798 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:5b7d3d7e5bdd0f7c2ad742990da2488e576cf0ea6dee2e3245192a89cc959096,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-75ff2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-cc776f956-cj2c7_openstack-operators(7d49c3c9-83c2-4c42-a318-c9d41ddf1bf8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:01:18 crc kubenswrapper[4934]: E1227 08:01:18.001754 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:1463c43243c75f56609cbae6bee2f86d411107181775721cb097cbd22fcae1d1" Dec 27 08:01:18 crc kubenswrapper[4934]: E1227 08:01:18.002207 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:1463c43243c75f56609cbae6bee2f86d411107181775721cb097cbd22fcae1d1,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jxhnw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-58879495c-d27qh_openstack-operators(89ea143d-f68d-46fc-b15a-6e00a418f65c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:01:20 crc kubenswrapper[4934]: E1227 08:01:20.117431 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.23:5001/openstack-k8s-operators/telemetry-operator:3e794ce9943831cfcd404a874792eef63230763c" Dec 27 08:01:20 crc kubenswrapper[4934]: E1227 08:01:20.117499 4934 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.23:5001/openstack-k8s-operators/telemetry-operator:3e794ce9943831cfcd404a874792eef63230763c" Dec 27 08:01:20 crc kubenswrapper[4934]: E1227 08:01:20.117661 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.23:5001/openstack-k8s-operators/telemetry-operator:3e794ce9943831cfcd404a874792eef63230763c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qjf9r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-6659cb75fc-9kzpj_openstack-operators(f2d5e345-1bd9-4313-bdae-7637cb7ed944): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:01:20 crc kubenswrapper[4934]: E1227 08:01:20.647884 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:b69a948d18394c4028a2957201d4dd83f17aea5dc07492fb797f871eeb0091df" Dec 27 08:01:20 crc kubenswrapper[4934]: E1227 08:01:20.648093 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:b69a948d18394c4028a2957201d4dd83f17aea5dc07492fb797f871eeb0091df,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qjdc5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-69977bdf55-p5jgx_openstack-operators(fc1c7052-d894-4fa1-b6bb-951cad9e32e0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:01:21 crc kubenswrapper[4934]: E1227 08:01:21.426266 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:d6a3d956e8dada1d7da372b532f955e6310002527667e24b08220c65956110af" Dec 27 08:01:21 crc kubenswrapper[4934]: E1227 08:01:21.426701 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:d6a3d956e8dada1d7da372b532f955e6310002527667e24b08220c65956110af,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6t7b8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-669b58f65-t89p4_openstack-operators(935a4c26-2678-4ead-87ee-1297850d89ff): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:01:22 crc kubenswrapper[4934]: E1227 08:01:22.653525 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 27 08:01:22 crc kubenswrapper[4934]: E1227 08:01:22.653776 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-77m2t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-z9x75_openstack-operators(e5e99871-c4b8-4bde-b7b7-41b7b5e2f5b0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:01:22 crc kubenswrapper[4934]: E1227 08:01:22.655717 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z9x75" podUID="e5e99871-c4b8-4bde-b7b7-41b7b5e2f5b0" Dec 27 08:01:23 crc kubenswrapper[4934]: E1227 08:01:23.030510 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:177bba84f71a0b2cfd00a31147aa349fe4c25c83d2b9df7563b5dd5cfeafc161" Dec 27 08:01:23 crc kubenswrapper[4934]: E1227 08:01:23.030683 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:177bba84f71a0b2cfd00a31147aa349fe4c25c83d2b9df7563b5dd5cfeafc161,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dfzj2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-6b444986fd-djrtj_openstack-operators(d090706d-34a7-4c72-9cb9-ad0601db35a6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:01:23 crc kubenswrapper[4934]: E1227 08:01:23.195881 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z9x75" podUID="e5e99871-c4b8-4bde-b7b7-41b7b5e2f5b0" Dec 27 08:01:23 crc kubenswrapper[4934]: E1227 08:01:23.545776 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" podUID="e2db190a-b455-44fc-a43e-2677e5af27b2" Dec 27 08:01:23 crc kubenswrapper[4934]: E1227 08:01:23.554016 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-58879495c-d27qh" podUID="89ea143d-f68d-46fc-b15a-6e00a418f65c" Dec 27 08:01:23 crc kubenswrapper[4934]: E1227 08:01:23.576517 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" podUID="59d65d09-c15c-49d9-8ee7-472ffe0dda55" Dec 27 08:01:23 crc kubenswrapper[4934]: E1227 08:01:23.592921 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-cc776f956-cj2c7" podUID="7d49c3c9-83c2-4c42-a318-c9d41ddf1bf8" Dec 27 08:01:23 crc kubenswrapper[4934]: E1227 08:01:23.665216 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg" podUID="4f825c5b-90ef-4b56-a759-d7ccbb312bed" Dec 27 08:01:23 crc kubenswrapper[4934]: E1227 08:01:23.675412 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59" podUID="828ebb68-326b-4c32-bc0a-d7f258b45ebd" Dec 27 08:01:23 crc kubenswrapper[4934]: E1227 08:01:23.732548 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" podUID="d090706d-34a7-4c72-9cb9-ad0601db35a6" Dec 27 08:01:23 crc kubenswrapper[4934]: E1227 08:01:23.762873 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj" podUID="f2d5e345-1bd9-4313-bdae-7637cb7ed944" Dec 27 08:01:23 crc kubenswrapper[4934]: E1227 08:01:23.788191 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4" podUID="935a4c26-2678-4ead-87ee-1297850d89ff" Dec 27 08:01:23 crc kubenswrapper[4934]: E1227 08:01:23.789387 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" podUID="f41746ba-5695-41bc-8078-29ba3d80e3f1" Dec 27 08:01:23 crc kubenswrapper[4934]: E1227 08:01:23.823697 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx" podUID="fc1c7052-d894-4fa1-b6bb-951cad9e32e0" Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.230858 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg" event={"ID":"4f825c5b-90ef-4b56-a759-d7ccbb312bed","Type":"ContainerStarted","Data":"e4ec38c65e881f732db7f00c769dec710f4cea597609136df041871e2c278b57"} Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.241361 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" event={"ID":"f41746ba-5695-41bc-8078-29ba3d80e3f1","Type":"ContainerStarted","Data":"22a5064e9e32cb4c60b0666d39cfb3e6a8c6101badcb8cafd9db69f5b9af7f48"} Dec 27 08:01:24 crc kubenswrapper[4934]: E1227 08:01:24.246805 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:317bf4faf2283cfc56db7596816bac63f7a6ecd5c7a83eec2ba0ea2f2b8c5448\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" podUID="f41746ba-5695-41bc-8078-29ba3d80e3f1" Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.287360 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cc599445b-dlwjl" event={"ID":"37d56061-eab6-43e5-8dec-f59220da5d0d","Type":"ContainerStarted","Data":"50262c84ec47924a26f26f43ac5a26fa2fc012c9e82081b4606dec8df1413648"} Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.307030 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" event={"ID":"e2db190a-b455-44fc-a43e-2677e5af27b2","Type":"ContainerStarted","Data":"f40e6a813b602571bd6e88ffac7b2034491998c02a214dc4362b8fc44e434cb5"} Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.309552 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-64fb555449-tbmpv" event={"ID":"3df60bce-0768-4949-8b8d-3bb28566cd4a","Type":"ContainerStarted","Data":"f4e8dd14b8a275c265dc0c74f505923d2e54467b74b7599d291217b628477431"} Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.313956 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" event={"ID":"35d6eb8d-3797-432c-a5ea-65f31adf7262","Type":"ContainerStarted","Data":"c637fdad1fae7799a7838d88477ef07194ee6d91bb834611b714d8bd903dab34"} Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.314683 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.327657 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" event={"ID":"59d65d09-c15c-49d9-8ee7-472ffe0dda55","Type":"ContainerStarted","Data":"d3d2b482b69b68dd8d4c72755843887794be0eb4a0662055da50355cea3c9723"} Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.339814 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" event={"ID":"77029090-da44-403b-a20f-d72105281956","Type":"ContainerStarted","Data":"f2c7f56b2f98c403b81d7d2c5df60467754fd6024eb74338cb3630cb95e8ca4b"} Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.347279 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4" event={"ID":"935a4c26-2678-4ead-87ee-1297850d89ff","Type":"ContainerStarted","Data":"dbbc00fec6751bed37bba594df4d42e33e82e0ae7e38799c27b91f0c4f67edc8"} Dec 27 08:01:24 crc kubenswrapper[4934]: E1227 08:01:24.352303 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:d6a3d956e8dada1d7da372b532f955e6310002527667e24b08220c65956110af\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4" podUID="935a4c26-2678-4ead-87ee-1297850d89ff" Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.361305 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" event={"ID":"d090706d-34a7-4c72-9cb9-ad0601db35a6","Type":"ContainerStarted","Data":"8afe1227da7de91de7f3930efe5e35c77b6a17054308fcea90a09cadccb8a8ce"} Dec 27 08:01:24 crc kubenswrapper[4934]: E1227 08:01:24.363997 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:177bba84f71a0b2cfd00a31147aa349fe4c25c83d2b9df7563b5dd5cfeafc161\\\"\"" pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" podUID="d090706d-34a7-4c72-9cb9-ad0601db35a6" Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.367166 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" event={"ID":"718122b6-c9f4-4d6f-92b5-6862c15f4ee6","Type":"ContainerStarted","Data":"ae6840f3f2f83c27940fbd80c73057b10eb4929c250ae9c6a06e465bba91b82c"} Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.376794 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-58879495c-d27qh" event={"ID":"89ea143d-f68d-46fc-b15a-6e00a418f65c","Type":"ContainerStarted","Data":"d818f8d8566befdae8c70490d946673653fef14fb1bb8fbce40b3f2bc7c3eb8c"} Dec 27 08:01:24 crc kubenswrapper[4934]: E1227 08:01:24.384971 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:1463c43243c75f56609cbae6bee2f86d411107181775721cb097cbd22fcae1d1\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-58879495c-d27qh" podUID="89ea143d-f68d-46fc-b15a-6e00a418f65c" Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.385175 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59" event={"ID":"828ebb68-326b-4c32-bc0a-d7f258b45ebd","Type":"ContainerStarted","Data":"6e610384d79930d4a7d84f19ed6fcf5083796f0bd6bc67897ebf23e8d541f0c7"} Dec 27 08:01:24 crc kubenswrapper[4934]: E1227 08:01:24.387272 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:2c4fe20e044dd8ea1f60f2f3f5e3844d932b4b79439835bd8771c73f16b38312\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59" podUID="828ebb68-326b-4c32-bc0a-d7f258b45ebd" Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.388538 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-cc776f956-cj2c7" event={"ID":"7d49c3c9-83c2-4c42-a318-c9d41ddf1bf8","Type":"ContainerStarted","Data":"5702587bc4c257e6146333b0584d1fb1c1b05e2ccf9deb3c7f1768018ad5e6ef"} Dec 27 08:01:24 crc kubenswrapper[4934]: E1227 08:01:24.389733 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:5b7d3d7e5bdd0f7c2ad742990da2488e576cf0ea6dee2e3245192a89cc959096\\\"\"" pod="openstack-operators/placement-operator-controller-manager-cc776f956-cj2c7" podUID="7d49c3c9-83c2-4c42-a318-c9d41ddf1bf8" Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.398650 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" event={"ID":"b5ea792c-b43f-4221-b7bb-aba3421de3d8","Type":"ContainerStarted","Data":"75dfa517cd5c84973d7563d3a3f58980a7e1672fd90a8b7a567fefc59a74e1f8"} Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.399647 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.415561 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" event={"ID":"1c0960f0-ef55-479a-b375-a09e69254743","Type":"ContainerStarted","Data":"9d102a6913460ad85934ce0bcd90edecc2fcb98d6472e06ef18cbbece71cf1bf"} Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.439018 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx" event={"ID":"fc1c7052-d894-4fa1-b6bb-951cad9e32e0","Type":"ContainerStarted","Data":"34953a9309c046b7b415268b42fe7ddab7fa4165738cb31c1d0c0c8736f2b54e"} Dec 27 08:01:24 crc kubenswrapper[4934]: E1227 08:01:24.444723 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:b69a948d18394c4028a2957201d4dd83f17aea5dc07492fb797f871eeb0091df\\\"\"" pod="openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx" podUID="fc1c7052-d894-4fa1-b6bb-951cad9e32e0" Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.448904 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr" event={"ID":"9fbceb4d-e7ae-42c0-8a3f-17807bc9f2ca","Type":"ContainerStarted","Data":"0fb7ff640bccf350e32ade99fb8f1d4e2ae0f966f789bd42b1ff56f9ef695f59"} Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.465954 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" podStartSLOduration=4.508056904 podStartE2EDuration="32.465936917s" podCreationTimestamp="2025-12-27 08:00:52 +0000 UTC" firstStartedPulling="2025-12-27 08:00:55.179543911 +0000 UTC m=+1115.999984505" lastFinishedPulling="2025-12-27 08:01:23.137423924 +0000 UTC m=+1143.957864518" observedRunningTime="2025-12-27 08:01:24.461830304 +0000 UTC m=+1145.282270898" watchObservedRunningTime="2025-12-27 08:01:24.465936917 +0000 UTC m=+1145.286377511" Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.468764 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj" event={"ID":"f2d5e345-1bd9-4313-bdae-7637cb7ed944","Type":"ContainerStarted","Data":"10027766d41293f3adac5c07c65a792a69119f6f509dc74f5c0565c04e95db0d"} Dec 27 08:01:24 crc kubenswrapper[4934]: E1227 08:01:24.470102 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.23:5001/openstack-k8s-operators/telemetry-operator:3e794ce9943831cfcd404a874792eef63230763c\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj" podUID="f2d5e345-1bd9-4313-bdae-7637cb7ed944" Dec 27 08:01:24 crc kubenswrapper[4934]: I1227 08:01:24.648579 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" podStartSLOduration=4.661186705 podStartE2EDuration="32.648564218s" podCreationTimestamp="2025-12-27 08:00:52 +0000 UTC" firstStartedPulling="2025-12-27 08:00:55.196931497 +0000 UTC m=+1116.017372091" lastFinishedPulling="2025-12-27 08:01:23.18430901 +0000 UTC m=+1144.004749604" observedRunningTime="2025-12-27 08:01:24.639941942 +0000 UTC m=+1145.460382536" watchObservedRunningTime="2025-12-27 08:01:24.648564218 +0000 UTC m=+1145.469004812" Dec 27 08:01:25 crc kubenswrapper[4934]: I1227 08:01:25.486466 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" event={"ID":"dc4ddef4-e9a5-4934-846e-1d1bcfe9f223","Type":"ContainerStarted","Data":"accb35096626791d61db24a47640ffc8de84625b32b0f51728a4076fcb71a160"} Dec 27 08:01:25 crc kubenswrapper[4934]: E1227 08:01:25.488550 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:1463c43243c75f56609cbae6bee2f86d411107181775721cb097cbd22fcae1d1\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-58879495c-d27qh" podUID="89ea143d-f68d-46fc-b15a-6e00a418f65c" Dec 27 08:01:25 crc kubenswrapper[4934]: E1227 08:01:25.488594 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:5b7d3d7e5bdd0f7c2ad742990da2488e576cf0ea6dee2e3245192a89cc959096\\\"\"" pod="openstack-operators/placement-operator-controller-manager-cc776f956-cj2c7" podUID="7d49c3c9-83c2-4c42-a318-c9d41ddf1bf8" Dec 27 08:01:25 crc kubenswrapper[4934]: E1227 08:01:25.488605 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:d6a3d956e8dada1d7da372b532f955e6310002527667e24b08220c65956110af\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4" podUID="935a4c26-2678-4ead-87ee-1297850d89ff" Dec 27 08:01:25 crc kubenswrapper[4934]: E1227 08:01:25.489513 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:b69a948d18394c4028a2957201d4dd83f17aea5dc07492fb797f871eeb0091df\\\"\"" pod="openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx" podUID="fc1c7052-d894-4fa1-b6bb-951cad9e32e0" Dec 27 08:01:25 crc kubenswrapper[4934]: E1227 08:01:25.489566 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.23:5001/openstack-k8s-operators/telemetry-operator:3e794ce9943831cfcd404a874792eef63230763c\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj" podUID="f2d5e345-1bd9-4313-bdae-7637cb7ed944" Dec 27 08:01:25 crc kubenswrapper[4934]: E1227 08:01:25.490549 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:177bba84f71a0b2cfd00a31147aa349fe4c25c83d2b9df7563b5dd5cfeafc161\\\"\"" pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" podUID="d090706d-34a7-4c72-9cb9-ad0601db35a6" Dec 27 08:01:25 crc kubenswrapper[4934]: E1227 08:01:25.490750 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:2c4fe20e044dd8ea1f60f2f3f5e3844d932b4b79439835bd8771c73f16b38312\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59" podUID="828ebb68-326b-4c32-bc0a-d7f258b45ebd" Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.525338 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cc599445b-dlwjl" event={"ID":"37d56061-eab6-43e5-8dec-f59220da5d0d","Type":"ContainerStarted","Data":"8ea9c5643250891eac4430b837f2a60e7e95a122a25f45461750519648f5f342"} Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.525786 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7cc599445b-dlwjl" Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.528342 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7cc599445b-dlwjl" Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.533338 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" event={"ID":"dc4ddef4-e9a5-4934-846e-1d1bcfe9f223","Type":"ContainerStarted","Data":"9639d5fa53404f1383bc02cbbf0595002ebaddffe39ab3d6326a370f0df86d23"} Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.534050 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.556146 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7cc599445b-dlwjl" podStartSLOduration=9.603136675 podStartE2EDuration="36.556125952s" podCreationTimestamp="2025-12-27 08:00:52 +0000 UTC" firstStartedPulling="2025-12-27 08:00:54.461202523 +0000 UTC m=+1115.281643117" lastFinishedPulling="2025-12-27 08:01:21.41419181 +0000 UTC m=+1142.234632394" observedRunningTime="2025-12-27 08:01:28.54568122 +0000 UTC m=+1149.366121814" watchObservedRunningTime="2025-12-27 08:01:28.556125952 +0000 UTC m=+1149.376566546" Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.563607 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr" event={"ID":"9fbceb4d-e7ae-42c0-8a3f-17807bc9f2ca","Type":"ContainerStarted","Data":"2835e9a844fbb59e4d8f8c655a8e34aa70e0581caa6839476da6d17aedf74e7d"} Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.563844 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr" Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.566708 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr" Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.578002 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-64fb555449-tbmpv" event={"ID":"3df60bce-0768-4949-8b8d-3bb28566cd4a","Type":"ContainerStarted","Data":"c8a116fac61cce7de830e670f9d6865d5b270dd7222c88104c1973a169347a33"} Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.579000 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-64fb555449-tbmpv" Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.593455 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-64fb555449-tbmpv" Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.615846 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" event={"ID":"1c0960f0-ef55-479a-b375-a09e69254743","Type":"ContainerStarted","Data":"da779ed0d3cfd6cfee26a1f79703dbd40fc0c586b501fbd804b88e0aea40bddb"} Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.616570 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.620815 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.623769 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" podStartSLOduration=8.327829846 podStartE2EDuration="36.623751239s" podCreationTimestamp="2025-12-27 08:00:52 +0000 UTC" firstStartedPulling="2025-12-27 08:00:53.587461376 +0000 UTC m=+1114.407901970" lastFinishedPulling="2025-12-27 08:01:21.883382769 +0000 UTC m=+1142.703823363" observedRunningTime="2025-12-27 08:01:28.611964593 +0000 UTC m=+1149.432405187" watchObservedRunningTime="2025-12-27 08:01:28.623751239 +0000 UTC m=+1149.444191833" Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.625553 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" event={"ID":"f1cce183-e0a8-4cf1-854e-7a60986f194f","Type":"ContainerStarted","Data":"6282fc1c681be3cc65189166b54cc594fee49777c42b330d5aae499cb8508585"} Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.626258 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.627483 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" event={"ID":"77029090-da44-403b-a20f-d72105281956","Type":"ContainerStarted","Data":"2c8833fd950bb70327670948ffed655c2832561c6742969fd2e3347f8e5fc989"} Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.628934 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.630166 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" event={"ID":"718122b6-c9f4-4d6f-92b5-6862c15f4ee6","Type":"ContainerStarted","Data":"2164ecc96286dd0ef8b42554f1f15074af9bdfe1cf3006dc2745b060b3f7b715"} Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.630782 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.630866 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.635881 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.647272 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr" podStartSLOduration=9.074646619 podStartE2EDuration="36.647240698s" podCreationTimestamp="2025-12-27 08:00:52 +0000 UTC" firstStartedPulling="2025-12-27 08:00:53.841589931 +0000 UTC m=+1114.662030525" lastFinishedPulling="2025-12-27 08:01:21.41418401 +0000 UTC m=+1142.234624604" observedRunningTime="2025-12-27 08:01:28.633206416 +0000 UTC m=+1149.453647020" watchObservedRunningTime="2025-12-27 08:01:28.647240698 +0000 UTC m=+1149.467681292" Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.700711 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" podStartSLOduration=9.227934733 podStartE2EDuration="36.700689458s" podCreationTimestamp="2025-12-27 08:00:52 +0000 UTC" firstStartedPulling="2025-12-27 08:00:53.94084924 +0000 UTC m=+1114.761289824" lastFinishedPulling="2025-12-27 08:01:21.413603955 +0000 UTC m=+1142.234044549" observedRunningTime="2025-12-27 08:01:28.697393656 +0000 UTC m=+1149.517834250" watchObservedRunningTime="2025-12-27 08:01:28.700689458 +0000 UTC m=+1149.521130052" Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.705567 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-64fb555449-tbmpv" podStartSLOduration=8.872023196 podStartE2EDuration="36.705549s" podCreationTimestamp="2025-12-27 08:00:52 +0000 UTC" firstStartedPulling="2025-12-27 08:00:53.578809739 +0000 UTC m=+1114.399250333" lastFinishedPulling="2025-12-27 08:01:21.412335543 +0000 UTC m=+1142.232776137" observedRunningTime="2025-12-27 08:01:28.671346602 +0000 UTC m=+1149.491787206" watchObservedRunningTime="2025-12-27 08:01:28.705549 +0000 UTC m=+1149.525989594" Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.731365 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" podStartSLOduration=8.766551572000001 podStartE2EDuration="36.731340737s" podCreationTimestamp="2025-12-27 08:00:52 +0000 UTC" firstStartedPulling="2025-12-27 08:00:55.219592786 +0000 UTC m=+1116.040033390" lastFinishedPulling="2025-12-27 08:01:23.184381971 +0000 UTC m=+1144.004822555" observedRunningTime="2025-12-27 08:01:28.720707991 +0000 UTC m=+1149.541148585" watchObservedRunningTime="2025-12-27 08:01:28.731340737 +0000 UTC m=+1149.551781341" Dec 27 08:01:28 crc kubenswrapper[4934]: I1227 08:01:28.794275 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" podStartSLOduration=9.200365261 podStartE2EDuration="36.794259575s" podCreationTimestamp="2025-12-27 08:00:52 +0000 UTC" firstStartedPulling="2025-12-27 08:00:53.819776263 +0000 UTC m=+1114.640216857" lastFinishedPulling="2025-12-27 08:01:21.413670587 +0000 UTC m=+1142.234111171" observedRunningTime="2025-12-27 08:01:28.753325709 +0000 UTC m=+1149.573766303" watchObservedRunningTime="2025-12-27 08:01:28.794259575 +0000 UTC m=+1149.614700169" Dec 27 08:01:29 crc kubenswrapper[4934]: I1227 08:01:29.641830 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg" event={"ID":"4f825c5b-90ef-4b56-a759-d7ccbb312bed","Type":"ContainerStarted","Data":"b343c2151f1146739a172297bf167ab133ea709a7937e7da93ca5ca5dfc7d0a0"} Dec 27 08:01:29 crc kubenswrapper[4934]: I1227 08:01:29.642149 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg" Dec 27 08:01:29 crc kubenswrapper[4934]: I1227 08:01:29.644718 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" event={"ID":"59d65d09-c15c-49d9-8ee7-472ffe0dda55","Type":"ContainerStarted","Data":"fd3b03e8c2268a330af22ea6886e6379543ac0c2538b1bc390292173c7cb0d81"} Dec 27 08:01:29 crc kubenswrapper[4934]: I1227 08:01:29.645536 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" Dec 27 08:01:29 crc kubenswrapper[4934]: I1227 08:01:29.647142 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" event={"ID":"f41746ba-5695-41bc-8078-29ba3d80e3f1","Type":"ContainerStarted","Data":"ff80d6933782e700e7230918f89286a2cea0af4dd10686c7b85ecbd92b9f6cb3"} Dec 27 08:01:29 crc kubenswrapper[4934]: I1227 08:01:29.647392 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" Dec 27 08:01:29 crc kubenswrapper[4934]: I1227 08:01:29.649186 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" event={"ID":"e2db190a-b455-44fc-a43e-2677e5af27b2","Type":"ContainerStarted","Data":"c9c80f7f97536e24c1f853c65d024959296423cac6628aefcc66960fe180b7d6"} Dec 27 08:01:29 crc kubenswrapper[4934]: I1227 08:01:29.653866 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" Dec 27 08:01:29 crc kubenswrapper[4934]: I1227 08:01:29.667063 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" podStartSLOduration=10.727266731 podStartE2EDuration="37.667041207s" podCreationTimestamp="2025-12-27 08:00:52 +0000 UTC" firstStartedPulling="2025-12-27 08:00:54.473345077 +0000 UTC m=+1115.293785671" lastFinishedPulling="2025-12-27 08:01:21.413119563 +0000 UTC m=+1142.233560147" observedRunningTime="2025-12-27 08:01:28.790374178 +0000 UTC m=+1149.610814772" watchObservedRunningTime="2025-12-27 08:01:29.667041207 +0000 UTC m=+1150.487481841" Dec 27 08:01:29 crc kubenswrapper[4934]: I1227 08:01:29.675409 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg" podStartSLOduration=4.554616292 podStartE2EDuration="37.675389036s" podCreationTimestamp="2025-12-27 08:00:52 +0000 UTC" firstStartedPulling="2025-12-27 08:00:55.162504294 +0000 UTC m=+1115.982944888" lastFinishedPulling="2025-12-27 08:01:28.283277018 +0000 UTC m=+1149.103717632" observedRunningTime="2025-12-27 08:01:29.663006646 +0000 UTC m=+1150.483447290" watchObservedRunningTime="2025-12-27 08:01:29.675389036 +0000 UTC m=+1150.495829670" Dec 27 08:01:29 crc kubenswrapper[4934]: I1227 08:01:29.732993 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" podStartSLOduration=3.702821306 podStartE2EDuration="37.732973471s" podCreationTimestamp="2025-12-27 08:00:52 +0000 UTC" firstStartedPulling="2025-12-27 08:00:54.400715155 +0000 UTC m=+1115.221155749" lastFinishedPulling="2025-12-27 08:01:28.43086732 +0000 UTC m=+1149.251307914" observedRunningTime="2025-12-27 08:01:29.725136314 +0000 UTC m=+1150.545576908" watchObservedRunningTime="2025-12-27 08:01:29.732973471 +0000 UTC m=+1150.553414085" Dec 27 08:01:29 crc kubenswrapper[4934]: I1227 08:01:29.759279 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" podStartSLOduration=3.9256659259999998 podStartE2EDuration="37.75926428s" podCreationTimestamp="2025-12-27 08:00:52 +0000 UTC" firstStartedPulling="2025-12-27 08:00:54.491346019 +0000 UTC m=+1115.311786603" lastFinishedPulling="2025-12-27 08:01:28.324944363 +0000 UTC m=+1149.145384957" observedRunningTime="2025-12-27 08:01:29.750531811 +0000 UTC m=+1150.570972405" watchObservedRunningTime="2025-12-27 08:01:29.75926428 +0000 UTC m=+1150.579704864" Dec 27 08:01:29 crc kubenswrapper[4934]: I1227 08:01:29.775250 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" podStartSLOduration=4.725452177 podStartE2EDuration="37.775224801s" podCreationTimestamp="2025-12-27 08:00:52 +0000 UTC" firstStartedPulling="2025-12-27 08:00:55.162565155 +0000 UTC m=+1115.983005749" lastFinishedPulling="2025-12-27 08:01:28.212337779 +0000 UTC m=+1149.032778373" observedRunningTime="2025-12-27 08:01:29.76843565 +0000 UTC m=+1150.588876284" watchObservedRunningTime="2025-12-27 08:01:29.775224801 +0000 UTC m=+1150.595665415" Dec 27 08:01:30 crc kubenswrapper[4934]: I1227 08:01:30.664806 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" Dec 27 08:01:33 crc kubenswrapper[4934]: I1227 08:01:33.092654 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" Dec 27 08:01:33 crc kubenswrapper[4934]: I1227 08:01:33.314283 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" Dec 27 08:01:33 crc kubenswrapper[4934]: I1227 08:01:33.339465 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" Dec 27 08:01:33 crc kubenswrapper[4934]: I1227 08:01:33.395354 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg" Dec 27 08:01:33 crc kubenswrapper[4934]: I1227 08:01:33.580032 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" Dec 27 08:01:33 crc kubenswrapper[4934]: I1227 08:01:33.605493 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" Dec 27 08:01:33 crc kubenswrapper[4934]: I1227 08:01:33.702117 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" Dec 27 08:01:38 crc kubenswrapper[4934]: I1227 08:01:38.752587 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" event={"ID":"d090706d-34a7-4c72-9cb9-ad0601db35a6","Type":"ContainerStarted","Data":"c498ab4a25c84f59bd3cc280f09132eeb7322b5390dd0cc7d9e2839317406180"} Dec 27 08:01:38 crc kubenswrapper[4934]: I1227 08:01:38.753808 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" Dec 27 08:01:38 crc kubenswrapper[4934]: I1227 08:01:38.754935 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj" event={"ID":"f2d5e345-1bd9-4313-bdae-7637cb7ed944","Type":"ContainerStarted","Data":"4d1d4d6175948e5f26e1a1bc49d40f0216381680eaf82c8d4d5a44a2bc8b57a3"} Dec 27 08:01:38 crc kubenswrapper[4934]: I1227 08:01:38.755195 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj" Dec 27 08:01:38 crc kubenswrapper[4934]: I1227 08:01:38.781787 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" podStartSLOduration=3.295686595 podStartE2EDuration="46.781764554s" podCreationTimestamp="2025-12-27 08:00:52 +0000 UTC" firstStartedPulling="2025-12-27 08:00:54.461166992 +0000 UTC m=+1115.281607586" lastFinishedPulling="2025-12-27 08:01:37.947244911 +0000 UTC m=+1158.767685545" observedRunningTime="2025-12-27 08:01:38.777099077 +0000 UTC m=+1159.597539681" watchObservedRunningTime="2025-12-27 08:01:38.781764554 +0000 UTC m=+1159.602205148" Dec 27 08:01:38 crc kubenswrapper[4934]: I1227 08:01:38.804706 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj" podStartSLOduration=4.4110402109999995 podStartE2EDuration="46.804682438s" podCreationTimestamp="2025-12-27 08:00:52 +0000 UTC" firstStartedPulling="2025-12-27 08:00:55.162166315 +0000 UTC m=+1115.982606909" lastFinishedPulling="2025-12-27 08:01:37.555808502 +0000 UTC m=+1158.376249136" observedRunningTime="2025-12-27 08:01:38.795529959 +0000 UTC m=+1159.615970553" watchObservedRunningTime="2025-12-27 08:01:38.804682438 +0000 UTC m=+1159.625123032" Dec 27 08:01:39 crc kubenswrapper[4934]: I1227 08:01:39.782991 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-cc776f956-cj2c7" event={"ID":"7d49c3c9-83c2-4c42-a318-c9d41ddf1bf8","Type":"ContainerStarted","Data":"43ac800bc63621ff01bad6299737651b39a52521251ff403eb96d70421b3ea02"} Dec 27 08:01:39 crc kubenswrapper[4934]: I1227 08:01:39.787308 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-cc776f956-cj2c7" Dec 27 08:01:39 crc kubenswrapper[4934]: I1227 08:01:39.796428 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4" event={"ID":"935a4c26-2678-4ead-87ee-1297850d89ff","Type":"ContainerStarted","Data":"a527cd3eae88fecad773add920717a3a4e06d78a3ae5704453a197e2e6f544e7"} Dec 27 08:01:39 crc kubenswrapper[4934]: I1227 08:01:39.796729 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4" Dec 27 08:01:39 crc kubenswrapper[4934]: I1227 08:01:39.801864 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z9x75" event={"ID":"e5e99871-c4b8-4bde-b7b7-41b7b5e2f5b0","Type":"ContainerStarted","Data":"64cc959d37e64ea0988b5f2193921386d5f43fd021d0179e6d9efa093c67a495"} Dec 27 08:01:39 crc kubenswrapper[4934]: I1227 08:01:39.815518 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-cc776f956-cj2c7" podStartSLOduration=3.821213546 podStartE2EDuration="47.815502283s" podCreationTimestamp="2025-12-27 08:00:52 +0000 UTC" firstStartedPulling="2025-12-27 08:00:55.162203276 +0000 UTC m=+1115.982643870" lastFinishedPulling="2025-12-27 08:01:39.156492023 +0000 UTC m=+1159.976932607" observedRunningTime="2025-12-27 08:01:39.811962355 +0000 UTC m=+1160.632402979" watchObservedRunningTime="2025-12-27 08:01:39.815502283 +0000 UTC m=+1160.635942877" Dec 27 08:01:39 crc kubenswrapper[4934]: I1227 08:01:39.831180 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4" podStartSLOduration=2.254684472 podStartE2EDuration="47.831160336s" podCreationTimestamp="2025-12-27 08:00:52 +0000 UTC" firstStartedPulling="2025-12-27 08:00:53.431947795 +0000 UTC m=+1114.252388389" lastFinishedPulling="2025-12-27 08:01:39.008423659 +0000 UTC m=+1159.828864253" observedRunningTime="2025-12-27 08:01:39.830747716 +0000 UTC m=+1160.651188340" watchObservedRunningTime="2025-12-27 08:01:39.831160336 +0000 UTC m=+1160.651600950" Dec 27 08:01:39 crc kubenswrapper[4934]: I1227 08:01:39.853671 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-z9x75" podStartSLOduration=3.085445143 podStartE2EDuration="46.85365365s" podCreationTimestamp="2025-12-27 08:00:53 +0000 UTC" firstStartedPulling="2025-12-27 08:00:55.236448548 +0000 UTC m=+1116.056889152" lastFinishedPulling="2025-12-27 08:01:39.004657065 +0000 UTC m=+1159.825097659" observedRunningTime="2025-12-27 08:01:39.848146652 +0000 UTC m=+1160.668587276" watchObservedRunningTime="2025-12-27 08:01:39.85365365 +0000 UTC m=+1160.674094244" Dec 27 08:01:40 crc kubenswrapper[4934]: I1227 08:01:40.810519 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59" event={"ID":"828ebb68-326b-4c32-bc0a-d7f258b45ebd","Type":"ContainerStarted","Data":"bc470b7b70a992d31e45ec11dbe0148fb6d26df5bc97a903ce0815d56b927ebb"} Dec 27 08:01:40 crc kubenswrapper[4934]: I1227 08:01:40.810760 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59" Dec 27 08:01:40 crc kubenswrapper[4934]: I1227 08:01:40.813124 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx" event={"ID":"fc1c7052-d894-4fa1-b6bb-951cad9e32e0","Type":"ContainerStarted","Data":"8db7241b4907307b83ebf5b5ab44035df5da16d52aa6adfd87254a21cf042bf4"} Dec 27 08:01:40 crc kubenswrapper[4934]: I1227 08:01:40.835783 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59" podStartSLOduration=3.127748232 podStartE2EDuration="48.835763194s" podCreationTimestamp="2025-12-27 08:00:52 +0000 UTC" firstStartedPulling="2025-12-27 08:00:54.303587649 +0000 UTC m=+1115.124028243" lastFinishedPulling="2025-12-27 08:01:40.011602611 +0000 UTC m=+1160.832043205" observedRunningTime="2025-12-27 08:01:40.832722208 +0000 UTC m=+1161.653162812" watchObservedRunningTime="2025-12-27 08:01:40.835763194 +0000 UTC m=+1161.656203778" Dec 27 08:01:40 crc kubenswrapper[4934]: I1227 08:01:40.864274 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx" podStartSLOduration=2.673767474 podStartE2EDuration="48.864256629s" podCreationTimestamp="2025-12-27 08:00:52 +0000 UTC" firstStartedPulling="2025-12-27 08:00:53.820214104 +0000 UTC m=+1114.640654698" lastFinishedPulling="2025-12-27 08:01:40.010703239 +0000 UTC m=+1160.831143853" observedRunningTime="2025-12-27 08:01:40.858110405 +0000 UTC m=+1161.678551039" watchObservedRunningTime="2025-12-27 08:01:40.864256629 +0000 UTC m=+1161.684697223" Dec 27 08:01:41 crc kubenswrapper[4934]: I1227 08:01:41.823727 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-58879495c-d27qh" event={"ID":"89ea143d-f68d-46fc-b15a-6e00a418f65c","Type":"ContainerStarted","Data":"6e5b6d1ae5ec9306af43b92ccbecbd76eebe0dee33dcdcfbf219295851b21002"} Dec 27 08:01:41 crc kubenswrapper[4934]: I1227 08:01:41.824681 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-58879495c-d27qh" Dec 27 08:01:41 crc kubenswrapper[4934]: I1227 08:01:41.856557 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-58879495c-d27qh" podStartSLOduration=3.322262581 podStartE2EDuration="49.856532659s" podCreationTimestamp="2025-12-27 08:00:52 +0000 UTC" firstStartedPulling="2025-12-27 08:00:54.395293999 +0000 UTC m=+1115.215734593" lastFinishedPulling="2025-12-27 08:01:40.929564077 +0000 UTC m=+1161.750004671" observedRunningTime="2025-12-27 08:01:41.84820326 +0000 UTC m=+1162.668643914" watchObservedRunningTime="2025-12-27 08:01:41.856532659 +0000 UTC m=+1162.676973283" Dec 27 08:01:42 crc kubenswrapper[4934]: I1227 08:01:42.641670 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx" Dec 27 08:01:43 crc kubenswrapper[4934]: I1227 08:01:43.232841 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" Dec 27 08:01:43 crc kubenswrapper[4934]: I1227 08:01:43.528096 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj" Dec 27 08:01:45 crc kubenswrapper[4934]: I1227 08:01:45.329808 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:01:45 crc kubenswrapper[4934]: I1227 08:01:45.330261 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:01:52 crc kubenswrapper[4934]: I1227 08:01:52.604415 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4" Dec 27 08:01:52 crc kubenswrapper[4934]: I1227 08:01:52.676751 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx" Dec 27 08:01:53 crc kubenswrapper[4934]: I1227 08:01:53.110691 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59" Dec 27 08:01:53 crc kubenswrapper[4934]: I1227 08:01:53.178602 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-58879495c-d27qh" Dec 27 08:01:53 crc kubenswrapper[4934]: I1227 08:01:53.335573 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-cc776f956-cj2c7" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.183119 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-rsz7l"] Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.187193 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-rsz7l" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.189963 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-pzjfs" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.190344 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.190485 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.190624 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.197355 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-rsz7l"] Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.255948 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-kmx4s"] Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.257637 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-kmx4s" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.259728 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.269530 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-kmx4s"] Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.379491 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74c5f0ab-04fb-4a95-a88a-8c448309e7fa-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-kmx4s\" (UID: \"74c5f0ab-04fb-4a95-a88a-8c448309e7fa\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kmx4s" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.379811 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3397fbf2-3f34-48bf-8c71-aa097ae68716-config\") pod \"dnsmasq-dns-675f4bcbfc-rsz7l\" (UID: \"3397fbf2-3f34-48bf-8c71-aa097ae68716\") " pod="openstack/dnsmasq-dns-675f4bcbfc-rsz7l" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.379937 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnfgw\" (UniqueName: \"kubernetes.io/projected/3397fbf2-3f34-48bf-8c71-aa097ae68716-kube-api-access-cnfgw\") pod \"dnsmasq-dns-675f4bcbfc-rsz7l\" (UID: \"3397fbf2-3f34-48bf-8c71-aa097ae68716\") " pod="openstack/dnsmasq-dns-675f4bcbfc-rsz7l" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.380213 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74c5f0ab-04fb-4a95-a88a-8c448309e7fa-config\") pod \"dnsmasq-dns-78dd6ddcc-kmx4s\" (UID: \"74c5f0ab-04fb-4a95-a88a-8c448309e7fa\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kmx4s" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.380345 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5knj2\" (UniqueName: \"kubernetes.io/projected/74c5f0ab-04fb-4a95-a88a-8c448309e7fa-kube-api-access-5knj2\") pod \"dnsmasq-dns-78dd6ddcc-kmx4s\" (UID: \"74c5f0ab-04fb-4a95-a88a-8c448309e7fa\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kmx4s" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.482374 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74c5f0ab-04fb-4a95-a88a-8c448309e7fa-config\") pod \"dnsmasq-dns-78dd6ddcc-kmx4s\" (UID: \"74c5f0ab-04fb-4a95-a88a-8c448309e7fa\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kmx4s" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.482434 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5knj2\" (UniqueName: \"kubernetes.io/projected/74c5f0ab-04fb-4a95-a88a-8c448309e7fa-kube-api-access-5knj2\") pod \"dnsmasq-dns-78dd6ddcc-kmx4s\" (UID: \"74c5f0ab-04fb-4a95-a88a-8c448309e7fa\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kmx4s" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.482516 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74c5f0ab-04fb-4a95-a88a-8c448309e7fa-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-kmx4s\" (UID: \"74c5f0ab-04fb-4a95-a88a-8c448309e7fa\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kmx4s" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.482540 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3397fbf2-3f34-48bf-8c71-aa097ae68716-config\") pod \"dnsmasq-dns-675f4bcbfc-rsz7l\" (UID: \"3397fbf2-3f34-48bf-8c71-aa097ae68716\") " pod="openstack/dnsmasq-dns-675f4bcbfc-rsz7l" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.482561 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnfgw\" (UniqueName: \"kubernetes.io/projected/3397fbf2-3f34-48bf-8c71-aa097ae68716-kube-api-access-cnfgw\") pod \"dnsmasq-dns-675f4bcbfc-rsz7l\" (UID: \"3397fbf2-3f34-48bf-8c71-aa097ae68716\") " pod="openstack/dnsmasq-dns-675f4bcbfc-rsz7l" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.483473 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74c5f0ab-04fb-4a95-a88a-8c448309e7fa-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-kmx4s\" (UID: \"74c5f0ab-04fb-4a95-a88a-8c448309e7fa\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kmx4s" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.483543 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74c5f0ab-04fb-4a95-a88a-8c448309e7fa-config\") pod \"dnsmasq-dns-78dd6ddcc-kmx4s\" (UID: \"74c5f0ab-04fb-4a95-a88a-8c448309e7fa\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kmx4s" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.483722 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3397fbf2-3f34-48bf-8c71-aa097ae68716-config\") pod \"dnsmasq-dns-675f4bcbfc-rsz7l\" (UID: \"3397fbf2-3f34-48bf-8c71-aa097ae68716\") " pod="openstack/dnsmasq-dns-675f4bcbfc-rsz7l" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.511445 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5knj2\" (UniqueName: \"kubernetes.io/projected/74c5f0ab-04fb-4a95-a88a-8c448309e7fa-kube-api-access-5knj2\") pod \"dnsmasq-dns-78dd6ddcc-kmx4s\" (UID: \"74c5f0ab-04fb-4a95-a88a-8c448309e7fa\") " pod="openstack/dnsmasq-dns-78dd6ddcc-kmx4s" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.512921 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnfgw\" (UniqueName: \"kubernetes.io/projected/3397fbf2-3f34-48bf-8c71-aa097ae68716-kube-api-access-cnfgw\") pod \"dnsmasq-dns-675f4bcbfc-rsz7l\" (UID: \"3397fbf2-3f34-48bf-8c71-aa097ae68716\") " pod="openstack/dnsmasq-dns-675f4bcbfc-rsz7l" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.578224 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-kmx4s" Dec 27 08:02:14 crc kubenswrapper[4934]: I1227 08:02:14.810210 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-rsz7l" Dec 27 08:02:15 crc kubenswrapper[4934]: I1227 08:02:15.029511 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-kmx4s"] Dec 27 08:02:15 crc kubenswrapper[4934]: I1227 08:02:15.202593 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-kmx4s" event={"ID":"74c5f0ab-04fb-4a95-a88a-8c448309e7fa","Type":"ContainerStarted","Data":"d3bcafe34da41a92388ef6ceded11435aac1a4133e28fd72c6cdcf04e2f85fd1"} Dec 27 08:02:15 crc kubenswrapper[4934]: I1227 08:02:15.280885 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-rsz7l"] Dec 27 08:02:15 crc kubenswrapper[4934]: I1227 08:02:15.330842 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:02:15 crc kubenswrapper[4934]: I1227 08:02:15.330901 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:02:15 crc kubenswrapper[4934]: I1227 08:02:15.330949 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 08:02:15 crc kubenswrapper[4934]: I1227 08:02:15.332583 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b960fbd3418f51dd41277ee06c40ec3eb700b9243e777f61c98b4b5b56bf0c98"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 08:02:15 crc kubenswrapper[4934]: I1227 08:02:15.332642 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://b960fbd3418f51dd41277ee06c40ec3eb700b9243e777f61c98b4b5b56bf0c98" gracePeriod=600 Dec 27 08:02:16 crc kubenswrapper[4934]: I1227 08:02:16.214302 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="b960fbd3418f51dd41277ee06c40ec3eb700b9243e777f61c98b4b5b56bf0c98" exitCode=0 Dec 27 08:02:16 crc kubenswrapper[4934]: I1227 08:02:16.214446 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"b960fbd3418f51dd41277ee06c40ec3eb700b9243e777f61c98b4b5b56bf0c98"} Dec 27 08:02:16 crc kubenswrapper[4934]: I1227 08:02:16.214757 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"b5285e4beb4b025eb5cca462bbc9a5da90660cedc2a9f99c3f12f8c689605503"} Dec 27 08:02:16 crc kubenswrapper[4934]: I1227 08:02:16.214777 4934 scope.go:117] "RemoveContainer" containerID="4ea115bb4d50c7db66f5a9c8846c893a8cb9f55d01ba76e6418fe5095edfd343" Dec 27 08:02:16 crc kubenswrapper[4934]: I1227 08:02:16.217203 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-rsz7l" event={"ID":"3397fbf2-3f34-48bf-8c71-aa097ae68716","Type":"ContainerStarted","Data":"2d987ecc491176028f4dfc538d1a7aa995a7aeff2d2b7d04ebce45a0e4a81a7f"} Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.053491 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-rsz7l"] Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.063622 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-6lbgv"] Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.066403 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.083221 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-6lbgv"] Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.235736 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff5d280b-73d2-49a6-bb58-4e515e88e179-config\") pod \"dnsmasq-dns-666b6646f7-6lbgv\" (UID: \"ff5d280b-73d2-49a6-bb58-4e515e88e179\") " pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.235795 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hczdl\" (UniqueName: \"kubernetes.io/projected/ff5d280b-73d2-49a6-bb58-4e515e88e179-kube-api-access-hczdl\") pod \"dnsmasq-dns-666b6646f7-6lbgv\" (UID: \"ff5d280b-73d2-49a6-bb58-4e515e88e179\") " pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.235841 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff5d280b-73d2-49a6-bb58-4e515e88e179-dns-svc\") pod \"dnsmasq-dns-666b6646f7-6lbgv\" (UID: \"ff5d280b-73d2-49a6-bb58-4e515e88e179\") " pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.305458 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-kmx4s"] Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.337965 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff5d280b-73d2-49a6-bb58-4e515e88e179-config\") pod \"dnsmasq-dns-666b6646f7-6lbgv\" (UID: \"ff5d280b-73d2-49a6-bb58-4e515e88e179\") " pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.338024 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hczdl\" (UniqueName: \"kubernetes.io/projected/ff5d280b-73d2-49a6-bb58-4e515e88e179-kube-api-access-hczdl\") pod \"dnsmasq-dns-666b6646f7-6lbgv\" (UID: \"ff5d280b-73d2-49a6-bb58-4e515e88e179\") " pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.338071 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff5d280b-73d2-49a6-bb58-4e515e88e179-dns-svc\") pod \"dnsmasq-dns-666b6646f7-6lbgv\" (UID: \"ff5d280b-73d2-49a6-bb58-4e515e88e179\") " pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.338876 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff5d280b-73d2-49a6-bb58-4e515e88e179-dns-svc\") pod \"dnsmasq-dns-666b6646f7-6lbgv\" (UID: \"ff5d280b-73d2-49a6-bb58-4e515e88e179\") " pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.339393 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff5d280b-73d2-49a6-bb58-4e515e88e179-config\") pod \"dnsmasq-dns-666b6646f7-6lbgv\" (UID: \"ff5d280b-73d2-49a6-bb58-4e515e88e179\") " pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.355724 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5fzqq"] Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.357399 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.364279 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hczdl\" (UniqueName: \"kubernetes.io/projected/ff5d280b-73d2-49a6-bb58-4e515e88e179-kube-api-access-hczdl\") pod \"dnsmasq-dns-666b6646f7-6lbgv\" (UID: \"ff5d280b-73d2-49a6-bb58-4e515e88e179\") " pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.364809 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5fzqq"] Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.392022 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.543697 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca9c1959-3f6a-4c4d-ab56-b464625d92c8-config\") pod \"dnsmasq-dns-57d769cc4f-5fzqq\" (UID: \"ca9c1959-3f6a-4c4d-ab56-b464625d92c8\") " pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.543764 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-955sp\" (UniqueName: \"kubernetes.io/projected/ca9c1959-3f6a-4c4d-ab56-b464625d92c8-kube-api-access-955sp\") pod \"dnsmasq-dns-57d769cc4f-5fzqq\" (UID: \"ca9c1959-3f6a-4c4d-ab56-b464625d92c8\") " pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.543881 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca9c1959-3f6a-4c4d-ab56-b464625d92c8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-5fzqq\" (UID: \"ca9c1959-3f6a-4c4d-ab56-b464625d92c8\") " pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.645410 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca9c1959-3f6a-4c4d-ab56-b464625d92c8-config\") pod \"dnsmasq-dns-57d769cc4f-5fzqq\" (UID: \"ca9c1959-3f6a-4c4d-ab56-b464625d92c8\") " pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.645776 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-955sp\" (UniqueName: \"kubernetes.io/projected/ca9c1959-3f6a-4c4d-ab56-b464625d92c8-kube-api-access-955sp\") pod \"dnsmasq-dns-57d769cc4f-5fzqq\" (UID: \"ca9c1959-3f6a-4c4d-ab56-b464625d92c8\") " pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.645931 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca9c1959-3f6a-4c4d-ab56-b464625d92c8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-5fzqq\" (UID: \"ca9c1959-3f6a-4c4d-ab56-b464625d92c8\") " pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.648637 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca9c1959-3f6a-4c4d-ab56-b464625d92c8-config\") pod \"dnsmasq-dns-57d769cc4f-5fzqq\" (UID: \"ca9c1959-3f6a-4c4d-ab56-b464625d92c8\") " pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.656018 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca9c1959-3f6a-4c4d-ab56-b464625d92c8-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-5fzqq\" (UID: \"ca9c1959-3f6a-4c4d-ab56-b464625d92c8\") " pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.687775 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-955sp\" (UniqueName: \"kubernetes.io/projected/ca9c1959-3f6a-4c4d-ab56-b464625d92c8-kube-api-access-955sp\") pod \"dnsmasq-dns-57d769cc4f-5fzqq\" (UID: \"ca9c1959-3f6a-4c4d-ab56-b464625d92c8\") " pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.722667 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" Dec 27 08:02:17 crc kubenswrapper[4934]: I1227 08:02:17.950659 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-6lbgv"] Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.177474 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.180196 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.189217 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.189561 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.189702 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.189811 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.192471 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.196309 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-drt5c" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.196581 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.196725 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.208198 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-1"] Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.209713 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.223045 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-2"] Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.224679 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.240480 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-1"] Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.257791 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-2"] Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.262238 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" event={"ID":"ff5d280b-73d2-49a6-bb58-4e515e88e179","Type":"ContainerStarted","Data":"e3ec4fc7e198727a2dd293607963b9e6f3a6f60a4a28ddec6396d7d2a06229c9"} Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.288559 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5fzqq"] Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360161 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360205 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360225 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5788f6c6-2500-4c66-af7a-830cde17caa4-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360254 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360278 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360299 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360329 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360455 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360482 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360508 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df7d6b06-cb87-4635-8aca-1ab00a564dd7-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360525 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr84v\" (UniqueName: \"kubernetes.io/projected/df7d6b06-cb87-4635-8aca-1ab00a564dd7-kube-api-access-qr84v\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360543 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5788f6c6-2500-4c66-af7a-830cde17caa4-server-conf\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360591 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360611 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360627 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df7d6b06-cb87-4635-8aca-1ab00a564dd7-pod-info\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360720 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-919d5745-3220-439b-a74b-98d67847c8df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-919d5745-3220-439b-a74b-98d67847c8df\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360776 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5788f6c6-2500-4c66-af7a-830cde17caa4-pod-info\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360806 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c561df02-e37e-4324-820c-98030eeb03e0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c561df02-e37e-4324-820c-98030eeb03e0\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360822 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360845 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8496x\" (UniqueName: \"kubernetes.io/projected/5788f6c6-2500-4c66-af7a-830cde17caa4-kube-api-access-8496x\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360878 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df7d6b06-cb87-4635-8aca-1ab00a564dd7-config-data\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360923 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.360986 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.361010 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-config-data\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.361054 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.361075 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5788f6c6-2500-4c66-af7a-830cde17caa4-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.361131 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.361160 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df7d6b06-cb87-4635-8aca-1ab00a564dd7-server-conf\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.361200 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5788f6c6-2500-4c66-af7a-830cde17caa4-config-data\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.361217 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s5mg\" (UniqueName: \"kubernetes.io/projected/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-kube-api-access-6s5mg\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.361236 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.361274 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.361293 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df7d6b06-cb87-4635-8aca-1ab00a564dd7-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.462493 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.462555 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.462592 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.462636 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.462689 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.462714 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.462735 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df7d6b06-cb87-4635-8aca-1ab00a564dd7-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.462753 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr84v\" (UniqueName: \"kubernetes.io/projected/df7d6b06-cb87-4635-8aca-1ab00a564dd7-kube-api-access-qr84v\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.462775 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5788f6c6-2500-4c66-af7a-830cde17caa4-server-conf\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.462805 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.462829 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.462850 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df7d6b06-cb87-4635-8aca-1ab00a564dd7-pod-info\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.462876 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-919d5745-3220-439b-a74b-98d67847c8df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-919d5745-3220-439b-a74b-98d67847c8df\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.462904 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5788f6c6-2500-4c66-af7a-830cde17caa4-pod-info\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.462929 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c561df02-e37e-4324-820c-98030eeb03e0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c561df02-e37e-4324-820c-98030eeb03e0\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.462953 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.462980 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8496x\" (UniqueName: \"kubernetes.io/projected/5788f6c6-2500-4c66-af7a-830cde17caa4-kube-api-access-8496x\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.463006 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df7d6b06-cb87-4635-8aca-1ab00a564dd7-config-data\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.463032 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.463058 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.463102 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-config-data\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.463134 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.463180 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5788f6c6-2500-4c66-af7a-830cde17caa4-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.463199 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.463218 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df7d6b06-cb87-4635-8aca-1ab00a564dd7-server-conf\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.463249 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s5mg\" (UniqueName: \"kubernetes.io/projected/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-kube-api-access-6s5mg\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.463264 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5788f6c6-2500-4c66-af7a-830cde17caa4-config-data\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.463284 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.463304 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.463320 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df7d6b06-cb87-4635-8aca-1ab00a564dd7-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.463340 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.463369 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.463385 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5788f6c6-2500-4c66-af7a-830cde17caa4-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.464246 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5788f6c6-2500-4c66-af7a-830cde17caa4-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.464643 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.464916 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df7d6b06-cb87-4635-8aca-1ab00a564dd7-config-data\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.464944 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5788f6c6-2500-4c66-af7a-830cde17caa4-server-conf\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.464957 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.465270 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.466398 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.466970 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.467242 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5788f6c6-2500-4c66-af7a-830cde17caa4-config-data\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.467751 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df7d6b06-cb87-4635-8aca-1ab00a564dd7-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.468997 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.469024 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/faea16fb88df6ea32d8bc003326a02149652e41cb41067608e5657db52571d30/globalmount\"" pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.471628 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.472540 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df7d6b06-cb87-4635-8aca-1ab00a564dd7-server-conf\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.473185 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-config-data\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.480441 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.502199 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df7d6b06-cb87-4635-8aca-1ab00a564dd7-pod-info\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.524565 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.525167 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.525727 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.526255 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.526730 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5788f6c6-2500-4c66-af7a-830cde17caa4-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.533741 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.534145 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.540933 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.552241 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.564558 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.566073 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.566140 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c561df02-e37e-4324-820c-98030eeb03e0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c561df02-e37e-4324-820c-98030eeb03e0\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5d9ca5747749969804eb38b0c8f9036813747c72b7012090db08a3a954d2ec6d/globalmount\"" pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.566691 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.566716 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-919d5745-3220-439b-a74b-98d67847c8df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-919d5745-3220-439b-a74b-98d67847c8df\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d634143c1674768228e9390fde62df5502f89e4584183c04f80432b07343da17/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.567230 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.572769 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.572992 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.573154 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.573273 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-kn2rk" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.575479 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.576251 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.594548 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5788f6c6-2500-4c66-af7a-830cde17caa4-pod-info\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.595755 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8496x\" (UniqueName: \"kubernetes.io/projected/5788f6c6-2500-4c66-af7a-830cde17caa4-kube-api-access-8496x\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.600750 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.601272 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.606407 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr84v\" (UniqueName: \"kubernetes.io/projected/df7d6b06-cb87-4635-8aca-1ab00a564dd7-kube-api-access-qr84v\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.611245 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df7d6b06-cb87-4635-8aca-1ab00a564dd7-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.630633 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s5mg\" (UniqueName: \"kubernetes.io/projected/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-kube-api-access-6s5mg\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.652236 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.719823 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/76158d63-2099-4418-82a2-09af0432c1a3-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.719871 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.719899 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.719961 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76158d63-2099-4418-82a2-09af0432c1a3-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.720053 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxr2j\" (UniqueName: \"kubernetes.io/projected/76158d63-2099-4418-82a2-09af0432c1a3-kube-api-access-lxr2j\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.720141 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/76158d63-2099-4418-82a2-09af0432c1a3-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.720235 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.720254 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/76158d63-2099-4418-82a2-09af0432c1a3-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.720305 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.720325 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/76158d63-2099-4418-82a2-09af0432c1a3-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.720366 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.779675 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c561df02-e37e-4324-820c-98030eeb03e0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c561df02-e37e-4324-820c-98030eeb03e0\") pod \"rabbitmq-server-2\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.791640 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-919d5745-3220-439b-a74b-98d67847c8df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-919d5745-3220-439b-a74b-98d67847c8df\") pod \"rabbitmq-server-0\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.801806 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2\") pod \"rabbitmq-server-1\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.807801 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.821870 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.821936 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/76158d63-2099-4418-82a2-09af0432c1a3-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.822607 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.823298 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.822374 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.823710 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.823784 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76158d63-2099-4418-82a2-09af0432c1a3-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.824899 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.824933 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/65d60d7c646e54e3493361ab222a8aba7e9f991db4c177aebfe00d23097f5c94/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.825363 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76158d63-2099-4418-82a2-09af0432c1a3-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.826278 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxr2j\" (UniqueName: \"kubernetes.io/projected/76158d63-2099-4418-82a2-09af0432c1a3-kube-api-access-lxr2j\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.826376 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/76158d63-2099-4418-82a2-09af0432c1a3-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.826485 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.826509 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/76158d63-2099-4418-82a2-09af0432c1a3-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.826569 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.826601 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/76158d63-2099-4418-82a2-09af0432c1a3-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.827610 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/76158d63-2099-4418-82a2-09af0432c1a3-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.827881 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/76158d63-2099-4418-82a2-09af0432c1a3-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.830991 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.831526 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/76158d63-2099-4418-82a2-09af0432c1a3-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.831000 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.834871 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/76158d63-2099-4418-82a2-09af0432c1a3-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.834968 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.849109 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.856456 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxr2j\" (UniqueName: \"kubernetes.io/projected/76158d63-2099-4418-82a2-09af0432c1a3-kube-api-access-lxr2j\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:18 crc kubenswrapper[4934]: I1227 08:02:18.879256 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e\") pod \"rabbitmq-cell1-server-0\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.043402 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.290059 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" event={"ID":"ca9c1959-3f6a-4c4d-ab56-b464625d92c8","Type":"ContainerStarted","Data":"cfcee839596b32cac9b5aeae2016f77d7ce8f5fe5896a08464822294fd807a87"} Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.313970 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 27 08:02:19 crc kubenswrapper[4934]: W1227 08:02:19.330920 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd9b61ad_7e6b_4ddd_b31b_b32ca0ad9a4c.slice/crio-3f2372d8ca89e99aac8a9574dbe4a130bc8c572cfa304e120ad6e5f99f44cc42 WatchSource:0}: Error finding container 3f2372d8ca89e99aac8a9574dbe4a130bc8c572cfa304e120ad6e5f99f44cc42: Status 404 returned error can't find the container with id 3f2372d8ca89e99aac8a9574dbe4a130bc8c572cfa304e120ad6e5f99f44cc42 Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.433249 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-2"] Dec 27 08:02:19 crc kubenswrapper[4934]: W1227 08:02:19.441836 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf7d6b06_cb87_4635_8aca_1ab00a564dd7.slice/crio-4f79111b4aad5ba928b46f818c525852ae66ed2784c724af6d61b660165a3f00 WatchSource:0}: Error finding container 4f79111b4aad5ba928b46f818c525852ae66ed2784c724af6d61b660165a3f00: Status 404 returned error can't find the container with id 4f79111b4aad5ba928b46f818c525852ae66ed2784c724af6d61b660165a3f00 Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.450248 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-1"] Dec 27 08:02:19 crc kubenswrapper[4934]: W1227 08:02:19.587453 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76158d63_2099_4418_82a2_09af0432c1a3.slice/crio-b8c7453eebadd180cef454914d2a20555fad716db7185abf26af4c945f82714d WatchSource:0}: Error finding container b8c7453eebadd180cef454914d2a20555fad716db7185abf26af4c945f82714d: Status 404 returned error can't find the container with id b8c7453eebadd180cef454914d2a20555fad716db7185abf26af4c945f82714d Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.594998 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.706731 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.709963 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.712989 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-87ztk" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.713235 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.713370 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.715509 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.716200 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.722837 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.726071 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.849255 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d03cf582-2192-4a10-b21e-10befebde419-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.849296 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d03cf582-2192-4a10-b21e-10befebde419-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.849343 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b4537546-dcf6-4a42-b2e7-5532633b40fa\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b4537546-dcf6-4a42-b2e7-5532633b40fa\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.849405 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d03cf582-2192-4a10-b21e-10befebde419-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.849525 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d03cf582-2192-4a10-b21e-10befebde419-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.849566 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fm4s\" (UniqueName: \"kubernetes.io/projected/d03cf582-2192-4a10-b21e-10befebde419-kube-api-access-7fm4s\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.849590 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d03cf582-2192-4a10-b21e-10befebde419-kolla-config\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.849629 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d03cf582-2192-4a10-b21e-10befebde419-secrets\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.849708 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d03cf582-2192-4a10-b21e-10befebde419-config-data-default\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.951364 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fm4s\" (UniqueName: \"kubernetes.io/projected/d03cf582-2192-4a10-b21e-10befebde419-kube-api-access-7fm4s\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.951415 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d03cf582-2192-4a10-b21e-10befebde419-kolla-config\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.951451 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d03cf582-2192-4a10-b21e-10befebde419-secrets\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.951520 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d03cf582-2192-4a10-b21e-10befebde419-config-data-default\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.951566 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d03cf582-2192-4a10-b21e-10befebde419-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.951583 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d03cf582-2192-4a10-b21e-10befebde419-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.951628 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d03cf582-2192-4a10-b21e-10befebde419-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.951646 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b4537546-dcf6-4a42-b2e7-5532633b40fa\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b4537546-dcf6-4a42-b2e7-5532633b40fa\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.951665 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d03cf582-2192-4a10-b21e-10befebde419-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.952363 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d03cf582-2192-4a10-b21e-10befebde419-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.952676 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d03cf582-2192-4a10-b21e-10befebde419-kolla-config\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.953421 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d03cf582-2192-4a10-b21e-10befebde419-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.953428 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d03cf582-2192-4a10-b21e-10befebde419-config-data-default\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.959497 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d03cf582-2192-4a10-b21e-10befebde419-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.960939 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.960973 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b4537546-dcf6-4a42-b2e7-5532633b40fa\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b4537546-dcf6-4a42-b2e7-5532633b40fa\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0ef828404aa020198ada9884a211d305a79c315c54d52328be29891d3126c846/globalmount\"" pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.961435 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d03cf582-2192-4a10-b21e-10befebde419-secrets\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.962309 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d03cf582-2192-4a10-b21e-10befebde419-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:19 crc kubenswrapper[4934]: I1227 08:02:19.977397 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fm4s\" (UniqueName: \"kubernetes.io/projected/d03cf582-2192-4a10-b21e-10befebde419-kube-api-access-7fm4s\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:20 crc kubenswrapper[4934]: I1227 08:02:20.002813 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b4537546-dcf6-4a42-b2e7-5532633b40fa\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b4537546-dcf6-4a42-b2e7-5532633b40fa\") pod \"openstack-galera-0\" (UID: \"d03cf582-2192-4a10-b21e-10befebde419\") " pod="openstack/openstack-galera-0" Dec 27 08:02:20 crc kubenswrapper[4934]: I1227 08:02:20.030693 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 27 08:02:20 crc kubenswrapper[4934]: I1227 08:02:20.310215 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"5788f6c6-2500-4c66-af7a-830cde17caa4","Type":"ContainerStarted","Data":"f4003fb78d58c40b4d21809dc8d97be99a25f219927cd511063ec4feb9101b6f"} Dec 27 08:02:20 crc kubenswrapper[4934]: I1227 08:02:20.312762 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"df7d6b06-cb87-4635-8aca-1ab00a564dd7","Type":"ContainerStarted","Data":"4f79111b4aad5ba928b46f818c525852ae66ed2784c724af6d61b660165a3f00"} Dec 27 08:02:20 crc kubenswrapper[4934]: I1227 08:02:20.315861 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"76158d63-2099-4418-82a2-09af0432c1a3","Type":"ContainerStarted","Data":"b8c7453eebadd180cef454914d2a20555fad716db7185abf26af4c945f82714d"} Dec 27 08:02:20 crc kubenswrapper[4934]: I1227 08:02:20.317483 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c","Type":"ContainerStarted","Data":"3f2372d8ca89e99aac8a9574dbe4a130bc8c572cfa304e120ad6e5f99f44cc42"} Dec 27 08:02:20 crc kubenswrapper[4934]: W1227 08:02:20.519740 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd03cf582_2192_4a10_b21e_10befebde419.slice/crio-07f513e575b8b8e1ff56a33935c04c68a100512309081ece1e6074527315ff54 WatchSource:0}: Error finding container 07f513e575b8b8e1ff56a33935c04c68a100512309081ece1e6074527315ff54: Status 404 returned error can't find the container with id 07f513e575b8b8e1ff56a33935c04c68a100512309081ece1e6074527315ff54 Dec 27 08:02:20 crc kubenswrapper[4934]: I1227 08:02:20.521506 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.036805 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.039741 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.045640 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.045800 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.045995 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.046285 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-mnrg5" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.054882 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.172987 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ca95a4ef-282d-4c26-983f-d7d00f90443c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.173032 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7bd7\" (UniqueName: \"kubernetes.io/projected/ca95a4ef-282d-4c26-983f-d7d00f90443c-kube-api-access-g7bd7\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.173069 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3a704175-dbf7-4a21-a749-d8cf6360f536\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a704175-dbf7-4a21-a749-d8cf6360f536\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.173113 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca95a4ef-282d-4c26-983f-d7d00f90443c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.173187 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ca95a4ef-282d-4c26-983f-d7d00f90443c-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.173207 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca95a4ef-282d-4c26-983f-d7d00f90443c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.173225 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ca95a4ef-282d-4c26-983f-d7d00f90443c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.173277 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ca95a4ef-282d-4c26-983f-d7d00f90443c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.173299 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca95a4ef-282d-4c26-983f-d7d00f90443c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.275281 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3a704175-dbf7-4a21-a749-d8cf6360f536\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a704175-dbf7-4a21-a749-d8cf6360f536\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.275346 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca95a4ef-282d-4c26-983f-d7d00f90443c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.275449 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ca95a4ef-282d-4c26-983f-d7d00f90443c-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.275476 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca95a4ef-282d-4c26-983f-d7d00f90443c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.275501 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ca95a4ef-282d-4c26-983f-d7d00f90443c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.275580 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ca95a4ef-282d-4c26-983f-d7d00f90443c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.276127 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ca95a4ef-282d-4c26-983f-d7d00f90443c-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.276212 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca95a4ef-282d-4c26-983f-d7d00f90443c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.276737 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ca95a4ef-282d-4c26-983f-d7d00f90443c-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.277333 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ca95a4ef-282d-4c26-983f-d7d00f90443c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.277375 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7bd7\" (UniqueName: \"kubernetes.io/projected/ca95a4ef-282d-4c26-983f-d7d00f90443c-kube-api-access-g7bd7\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.277394 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca95a4ef-282d-4c26-983f-d7d00f90443c-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.277780 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ca95a4ef-282d-4c26-983f-d7d00f90443c-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.278618 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.278645 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3a704175-dbf7-4a21-a749-d8cf6360f536\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a704175-dbf7-4a21-a749-d8cf6360f536\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/7512a407658b9b1c5bdc0941bb248bb5fff097845a466a065e4b4c91dfe219c6/globalmount\"" pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.282454 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca95a4ef-282d-4c26-983f-d7d00f90443c-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.282727 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca95a4ef-282d-4c26-983f-d7d00f90443c-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.293260 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ca95a4ef-282d-4c26-983f-d7d00f90443c-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.307484 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7bd7\" (UniqueName: \"kubernetes.io/projected/ca95a4ef-282d-4c26-983f-d7d00f90443c-kube-api-access-g7bd7\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.330979 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3a704175-dbf7-4a21-a749-d8cf6360f536\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3a704175-dbf7-4a21-a749-d8cf6360f536\") pod \"openstack-cell1-galera-0\" (UID: \"ca95a4ef-282d-4c26-983f-d7d00f90443c\") " pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.335708 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d03cf582-2192-4a10-b21e-10befebde419","Type":"ContainerStarted","Data":"07f513e575b8b8e1ff56a33935c04c68a100512309081ece1e6074527315ff54"} Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.370908 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.443407 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.448909 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.452545 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.452690 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-pnbj8" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.453053 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.488472 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.582873 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b102f157-6ec2-4b75-846d-8362a74bf9d2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"b102f157-6ec2-4b75-846d-8362a74bf9d2\") " pod="openstack/memcached-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.582998 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b102f157-6ec2-4b75-846d-8362a74bf9d2-config-data\") pod \"memcached-0\" (UID: \"b102f157-6ec2-4b75-846d-8362a74bf9d2\") " pod="openstack/memcached-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.583059 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b102f157-6ec2-4b75-846d-8362a74bf9d2-combined-ca-bundle\") pod \"memcached-0\" (UID: \"b102f157-6ec2-4b75-846d-8362a74bf9d2\") " pod="openstack/memcached-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.583146 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7mdd\" (UniqueName: \"kubernetes.io/projected/b102f157-6ec2-4b75-846d-8362a74bf9d2-kube-api-access-p7mdd\") pod \"memcached-0\" (UID: \"b102f157-6ec2-4b75-846d-8362a74bf9d2\") " pod="openstack/memcached-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.583176 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b102f157-6ec2-4b75-846d-8362a74bf9d2-kolla-config\") pod \"memcached-0\" (UID: \"b102f157-6ec2-4b75-846d-8362a74bf9d2\") " pod="openstack/memcached-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.684194 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b102f157-6ec2-4b75-846d-8362a74bf9d2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"b102f157-6ec2-4b75-846d-8362a74bf9d2\") " pod="openstack/memcached-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.684268 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b102f157-6ec2-4b75-846d-8362a74bf9d2-config-data\") pod \"memcached-0\" (UID: \"b102f157-6ec2-4b75-846d-8362a74bf9d2\") " pod="openstack/memcached-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.684313 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b102f157-6ec2-4b75-846d-8362a74bf9d2-combined-ca-bundle\") pod \"memcached-0\" (UID: \"b102f157-6ec2-4b75-846d-8362a74bf9d2\") " pod="openstack/memcached-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.684353 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7mdd\" (UniqueName: \"kubernetes.io/projected/b102f157-6ec2-4b75-846d-8362a74bf9d2-kube-api-access-p7mdd\") pod \"memcached-0\" (UID: \"b102f157-6ec2-4b75-846d-8362a74bf9d2\") " pod="openstack/memcached-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.684379 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b102f157-6ec2-4b75-846d-8362a74bf9d2-kolla-config\") pod \"memcached-0\" (UID: \"b102f157-6ec2-4b75-846d-8362a74bf9d2\") " pod="openstack/memcached-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.685070 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b102f157-6ec2-4b75-846d-8362a74bf9d2-kolla-config\") pod \"memcached-0\" (UID: \"b102f157-6ec2-4b75-846d-8362a74bf9d2\") " pod="openstack/memcached-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.685338 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b102f157-6ec2-4b75-846d-8362a74bf9d2-config-data\") pod \"memcached-0\" (UID: \"b102f157-6ec2-4b75-846d-8362a74bf9d2\") " pod="openstack/memcached-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.688698 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b102f157-6ec2-4b75-846d-8362a74bf9d2-combined-ca-bundle\") pod \"memcached-0\" (UID: \"b102f157-6ec2-4b75-846d-8362a74bf9d2\") " pod="openstack/memcached-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.700504 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/b102f157-6ec2-4b75-846d-8362a74bf9d2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"b102f157-6ec2-4b75-846d-8362a74bf9d2\") " pod="openstack/memcached-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.702246 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7mdd\" (UniqueName: \"kubernetes.io/projected/b102f157-6ec2-4b75-846d-8362a74bf9d2-kube-api-access-p7mdd\") pod \"memcached-0\" (UID: \"b102f157-6ec2-4b75-846d-8362a74bf9d2\") " pod="openstack/memcached-0" Dec 27 08:02:21 crc kubenswrapper[4934]: I1227 08:02:21.767844 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 27 08:02:23 crc kubenswrapper[4934]: I1227 08:02:23.486268 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 27 08:02:23 crc kubenswrapper[4934]: I1227 08:02:23.504689 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 27 08:02:23 crc kubenswrapper[4934]: I1227 08:02:23.504699 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 27 08:02:23 crc kubenswrapper[4934]: I1227 08:02:23.507966 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-6kpjb" Dec 27 08:02:23 crc kubenswrapper[4934]: I1227 08:02:23.624211 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk66k\" (UniqueName: \"kubernetes.io/projected/26cea6d5-e2df-46ba-a731-722d91bd3c91-kube-api-access-nk66k\") pod \"kube-state-metrics-0\" (UID: \"26cea6d5-e2df-46ba-a731-722d91bd3c91\") " pod="openstack/kube-state-metrics-0" Dec 27 08:02:23 crc kubenswrapper[4934]: I1227 08:02:23.727459 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk66k\" (UniqueName: \"kubernetes.io/projected/26cea6d5-e2df-46ba-a731-722d91bd3c91-kube-api-access-nk66k\") pod \"kube-state-metrics-0\" (UID: \"26cea6d5-e2df-46ba-a731-722d91bd3c91\") " pod="openstack/kube-state-metrics-0" Dec 27 08:02:23 crc kubenswrapper[4934]: I1227 08:02:23.776843 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk66k\" (UniqueName: \"kubernetes.io/projected/26cea6d5-e2df-46ba-a731-722d91bd3c91-kube-api-access-nk66k\") pod \"kube-state-metrics-0\" (UID: \"26cea6d5-e2df-46ba-a731-722d91bd3c91\") " pod="openstack/kube-state-metrics-0" Dec 27 08:02:23 crc kubenswrapper[4934]: I1227 08:02:23.830518 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.324046 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-ui-dashboards-66cbf594b5-swd4w"] Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.327487 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-66cbf594b5-swd4w" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.329855 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.330123 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards-sa-dockercfg-7v5cd" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.332500 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-66cbf594b5-swd4w"] Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.450700 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/355bcd5e-7852-477f-8e4b-b03847566cc6-serving-cert\") pod \"observability-ui-dashboards-66cbf594b5-swd4w\" (UID: \"355bcd5e-7852-477f-8e4b-b03847566cc6\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-swd4w" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.451285 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhrvn\" (UniqueName: \"kubernetes.io/projected/355bcd5e-7852-477f-8e4b-b03847566cc6-kube-api-access-zhrvn\") pod \"observability-ui-dashboards-66cbf594b5-swd4w\" (UID: \"355bcd5e-7852-477f-8e4b-b03847566cc6\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-swd4w" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.554027 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhrvn\" (UniqueName: \"kubernetes.io/projected/355bcd5e-7852-477f-8e4b-b03847566cc6-kube-api-access-zhrvn\") pod \"observability-ui-dashboards-66cbf594b5-swd4w\" (UID: \"355bcd5e-7852-477f-8e4b-b03847566cc6\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-swd4w" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.554182 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/355bcd5e-7852-477f-8e4b-b03847566cc6-serving-cert\") pod \"observability-ui-dashboards-66cbf594b5-swd4w\" (UID: \"355bcd5e-7852-477f-8e4b-b03847566cc6\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-swd4w" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.562711 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/355bcd5e-7852-477f-8e4b-b03847566cc6-serving-cert\") pod \"observability-ui-dashboards-66cbf594b5-swd4w\" (UID: \"355bcd5e-7852-477f-8e4b-b03847566cc6\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-swd4w" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.578657 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhrvn\" (UniqueName: \"kubernetes.io/projected/355bcd5e-7852-477f-8e4b-b03847566cc6-kube-api-access-zhrvn\") pod \"observability-ui-dashboards-66cbf594b5-swd4w\" (UID: \"355bcd5e-7852-477f-8e4b-b03847566cc6\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-swd4w" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.641209 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-74cc7c486-2dpkp"] Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.642521 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.654938 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-74cc7c486-2dpkp"] Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.663549 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-66cbf594b5-swd4w" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.758614 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e0978ad2-1582-49d4-b239-ff929dafb117-oauth-serving-cert\") pod \"console-74cc7c486-2dpkp\" (UID: \"e0978ad2-1582-49d4-b239-ff929dafb117\") " pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.758782 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e0978ad2-1582-49d4-b239-ff929dafb117-console-serving-cert\") pod \"console-74cc7c486-2dpkp\" (UID: \"e0978ad2-1582-49d4-b239-ff929dafb117\") " pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.758860 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e0978ad2-1582-49d4-b239-ff929dafb117-console-oauth-config\") pod \"console-74cc7c486-2dpkp\" (UID: \"e0978ad2-1582-49d4-b239-ff929dafb117\") " pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.759027 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e0978ad2-1582-49d4-b239-ff929dafb117-trusted-ca-bundle\") pod \"console-74cc7c486-2dpkp\" (UID: \"e0978ad2-1582-49d4-b239-ff929dafb117\") " pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.759136 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e0978ad2-1582-49d4-b239-ff929dafb117-service-ca\") pod \"console-74cc7c486-2dpkp\" (UID: \"e0978ad2-1582-49d4-b239-ff929dafb117\") " pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.759197 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sckk\" (UniqueName: \"kubernetes.io/projected/e0978ad2-1582-49d4-b239-ff929dafb117-kube-api-access-2sckk\") pod \"console-74cc7c486-2dpkp\" (UID: \"e0978ad2-1582-49d4-b239-ff929dafb117\") " pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.759227 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e0978ad2-1582-49d4-b239-ff929dafb117-console-config\") pod \"console-74cc7c486-2dpkp\" (UID: \"e0978ad2-1582-49d4-b239-ff929dafb117\") " pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.839774 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.842798 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.848112 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.858559 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.858701 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.858818 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.858963 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-9q2gk" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.859146 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.859281 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-2" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.859405 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-1" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.860542 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e0978ad2-1582-49d4-b239-ff929dafb117-console-config\") pod \"console-74cc7c486-2dpkp\" (UID: \"e0978ad2-1582-49d4-b239-ff929dafb117\") " pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.860564 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e0978ad2-1582-49d4-b239-ff929dafb117-service-ca\") pod \"console-74cc7c486-2dpkp\" (UID: \"e0978ad2-1582-49d4-b239-ff929dafb117\") " pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.860581 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sckk\" (UniqueName: \"kubernetes.io/projected/e0978ad2-1582-49d4-b239-ff929dafb117-kube-api-access-2sckk\") pod \"console-74cc7c486-2dpkp\" (UID: \"e0978ad2-1582-49d4-b239-ff929dafb117\") " pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.860638 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e0978ad2-1582-49d4-b239-ff929dafb117-oauth-serving-cert\") pod \"console-74cc7c486-2dpkp\" (UID: \"e0978ad2-1582-49d4-b239-ff929dafb117\") " pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.860703 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e0978ad2-1582-49d4-b239-ff929dafb117-console-serving-cert\") pod \"console-74cc7c486-2dpkp\" (UID: \"e0978ad2-1582-49d4-b239-ff929dafb117\") " pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.860743 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e0978ad2-1582-49d4-b239-ff929dafb117-console-oauth-config\") pod \"console-74cc7c486-2dpkp\" (UID: \"e0978ad2-1582-49d4-b239-ff929dafb117\") " pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.860811 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e0978ad2-1582-49d4-b239-ff929dafb117-trusted-ca-bundle\") pod \"console-74cc7c486-2dpkp\" (UID: \"e0978ad2-1582-49d4-b239-ff929dafb117\") " pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.861993 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e0978ad2-1582-49d4-b239-ff929dafb117-oauth-serving-cert\") pod \"console-74cc7c486-2dpkp\" (UID: \"e0978ad2-1582-49d4-b239-ff929dafb117\") " pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.862428 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e0978ad2-1582-49d4-b239-ff929dafb117-console-config\") pod \"console-74cc7c486-2dpkp\" (UID: \"e0978ad2-1582-49d4-b239-ff929dafb117\") " pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.862443 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e0978ad2-1582-49d4-b239-ff929dafb117-service-ca\") pod \"console-74cc7c486-2dpkp\" (UID: \"e0978ad2-1582-49d4-b239-ff929dafb117\") " pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.863075 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.864872 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e0978ad2-1582-49d4-b239-ff929dafb117-trusted-ca-bundle\") pod \"console-74cc7c486-2dpkp\" (UID: \"e0978ad2-1582-49d4-b239-ff929dafb117\") " pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.866942 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e0978ad2-1582-49d4-b239-ff929dafb117-console-oauth-config\") pod \"console-74cc7c486-2dpkp\" (UID: \"e0978ad2-1582-49d4-b239-ff929dafb117\") " pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.867457 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e0978ad2-1582-49d4-b239-ff929dafb117-console-serving-cert\") pod \"console-74cc7c486-2dpkp\" (UID: \"e0978ad2-1582-49d4-b239-ff929dafb117\") " pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.883350 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sckk\" (UniqueName: \"kubernetes.io/projected/e0978ad2-1582-49d4-b239-ff929dafb117-kube-api-access-2sckk\") pod \"console-74cc7c486-2dpkp\" (UID: \"e0978ad2-1582-49d4-b239-ff929dafb117\") " pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.976243 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92d66221-653f-4837-a733-f02073b6ed70-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.976416 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vr285\" (UniqueName: \"kubernetes.io/projected/92d66221-653f-4837-a733-f02073b6ed70-kube-api-access-vr285\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.976466 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/92d66221-653f-4837-a733-f02073b6ed70-config\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.976527 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/92d66221-653f-4837-a733-f02073b6ed70-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.976573 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92d66221-653f-4837-a733-f02073b6ed70-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.976593 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92d66221-653f-4837-a733-f02073b6ed70-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.976657 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92d66221-653f-4837-a733-f02073b6ed70-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.976679 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92d66221-653f-4837-a733-f02073b6ed70-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.976715 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/92d66221-653f-4837-a733-f02073b6ed70-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.976763 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:24 crc kubenswrapper[4934]: I1227 08:02:24.983670 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.079520 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vr285\" (UniqueName: \"kubernetes.io/projected/92d66221-653f-4837-a733-f02073b6ed70-kube-api-access-vr285\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.079618 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/92d66221-653f-4837-a733-f02073b6ed70-config\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.079701 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/92d66221-653f-4837-a733-f02073b6ed70-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.079778 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92d66221-653f-4837-a733-f02073b6ed70-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.079823 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92d66221-653f-4837-a733-f02073b6ed70-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.079871 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92d66221-653f-4837-a733-f02073b6ed70-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.079911 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92d66221-653f-4837-a733-f02073b6ed70-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.079941 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/92d66221-653f-4837-a733-f02073b6ed70-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.080104 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.080154 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92d66221-653f-4837-a733-f02073b6ed70-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.081123 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92d66221-653f-4837-a733-f02073b6ed70-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.082301 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/92d66221-653f-4837-a733-f02073b6ed70-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.082817 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/92d66221-653f-4837-a733-f02073b6ed70-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.083891 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92d66221-653f-4837-a733-f02073b6ed70-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.085301 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92d66221-653f-4837-a733-f02073b6ed70-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.085686 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92d66221-653f-4837-a733-f02073b6ed70-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.088452 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92d66221-653f-4837-a733-f02073b6ed70-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.093169 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.093358 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b43fafdc1175ed3c7018bef2bd4eb270b5f0d66d6e72f88cdbf5483bdbd6cc4f/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.099678 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vr285\" (UniqueName: \"kubernetes.io/projected/92d66221-653f-4837-a733-f02073b6ed70-kube-api-access-vr285\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.108725 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/92d66221-653f-4837-a733-f02073b6ed70-config\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.156950 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b\") pod \"prometheus-metric-storage-0\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:25 crc kubenswrapper[4934]: I1227 08:02:25.271848 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.395873 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-22dwz"] Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.397378 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.406189 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-5qvx6" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.406409 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.406585 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.411201 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-22dwz"] Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.422893 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-bdjjv"] Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.425147 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.430782 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-bdjjv"] Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.481555 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.483034 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.491863 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.492008 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-knrs6" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.492164 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.492240 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.492300 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.497480 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.524362 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f4b9eef2-987b-4c54-803c-6655aceab8f6-var-log-ovn\") pod \"ovn-controller-22dwz\" (UID: \"f4b9eef2-987b-4c54-803c-6655aceab8f6\") " pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.524397 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b9eef2-987b-4c54-803c-6655aceab8f6-combined-ca-bundle\") pod \"ovn-controller-22dwz\" (UID: \"f4b9eef2-987b-4c54-803c-6655aceab8f6\") " pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.524471 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/54ab7f4e-f623-473b-8b6a-3363fe487b4e-var-run\") pod \"ovn-controller-ovs-bdjjv\" (UID: \"54ab7f4e-f623-473b-8b6a-3363fe487b4e\") " pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.524679 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f4b9eef2-987b-4c54-803c-6655aceab8f6-var-run\") pod \"ovn-controller-22dwz\" (UID: \"f4b9eef2-987b-4c54-803c-6655aceab8f6\") " pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.524708 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4b9eef2-987b-4c54-803c-6655aceab8f6-scripts\") pod \"ovn-controller-22dwz\" (UID: \"f4b9eef2-987b-4c54-803c-6655aceab8f6\") " pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.524745 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m9jp\" (UniqueName: \"kubernetes.io/projected/54ab7f4e-f623-473b-8b6a-3363fe487b4e-kube-api-access-7m9jp\") pod \"ovn-controller-ovs-bdjjv\" (UID: \"54ab7f4e-f623-473b-8b6a-3363fe487b4e\") " pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.524776 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzh6b\" (UniqueName: \"kubernetes.io/projected/f4b9eef2-987b-4c54-803c-6655aceab8f6-kube-api-access-fzh6b\") pod \"ovn-controller-22dwz\" (UID: \"f4b9eef2-987b-4c54-803c-6655aceab8f6\") " pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.524806 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f4b9eef2-987b-4c54-803c-6655aceab8f6-var-run-ovn\") pod \"ovn-controller-22dwz\" (UID: \"f4b9eef2-987b-4c54-803c-6655aceab8f6\") " pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.525552 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/54ab7f4e-f623-473b-8b6a-3363fe487b4e-etc-ovs\") pod \"ovn-controller-ovs-bdjjv\" (UID: \"54ab7f4e-f623-473b-8b6a-3363fe487b4e\") " pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.525581 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/54ab7f4e-f623-473b-8b6a-3363fe487b4e-var-lib\") pod \"ovn-controller-ovs-bdjjv\" (UID: \"54ab7f4e-f623-473b-8b6a-3363fe487b4e\") " pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.525595 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54ab7f4e-f623-473b-8b6a-3363fe487b4e-scripts\") pod \"ovn-controller-ovs-bdjjv\" (UID: \"54ab7f4e-f623-473b-8b6a-3363fe487b4e\") " pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.525611 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4b9eef2-987b-4c54-803c-6655aceab8f6-ovn-controller-tls-certs\") pod \"ovn-controller-22dwz\" (UID: \"f4b9eef2-987b-4c54-803c-6655aceab8f6\") " pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.525655 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/54ab7f4e-f623-473b-8b6a-3363fe487b4e-var-log\") pod \"ovn-controller-ovs-bdjjv\" (UID: \"54ab7f4e-f623-473b-8b6a-3363fe487b4e\") " pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.627308 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc83d812-aa8d-45e4-b6df-02e6f8a9d9db-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.627366 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/54ab7f4e-f623-473b-8b6a-3363fe487b4e-var-run\") pod \"ovn-controller-ovs-bdjjv\" (UID: \"54ab7f4e-f623-473b-8b6a-3363fe487b4e\") " pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.627425 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6c3d800f-2362-434e-a5ce-3e291bfdcf14\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6c3d800f-2362-434e-a5ce-3e291bfdcf14\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.627458 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f4b9eef2-987b-4c54-803c-6655aceab8f6-var-run\") pod \"ovn-controller-22dwz\" (UID: \"f4b9eef2-987b-4c54-803c-6655aceab8f6\") " pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.627483 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4b9eef2-987b-4c54-803c-6655aceab8f6-scripts\") pod \"ovn-controller-22dwz\" (UID: \"f4b9eef2-987b-4c54-803c-6655aceab8f6\") " pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.627905 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/54ab7f4e-f623-473b-8b6a-3363fe487b4e-var-run\") pod \"ovn-controller-ovs-bdjjv\" (UID: \"54ab7f4e-f623-473b-8b6a-3363fe487b4e\") " pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.628153 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f4b9eef2-987b-4c54-803c-6655aceab8f6-var-run\") pod \"ovn-controller-22dwz\" (UID: \"f4b9eef2-987b-4c54-803c-6655aceab8f6\") " pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.631958 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m9jp\" (UniqueName: \"kubernetes.io/projected/54ab7f4e-f623-473b-8b6a-3363fe487b4e-kube-api-access-7m9jp\") pod \"ovn-controller-ovs-bdjjv\" (UID: \"54ab7f4e-f623-473b-8b6a-3363fe487b4e\") " pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.632579 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzh6b\" (UniqueName: \"kubernetes.io/projected/f4b9eef2-987b-4c54-803c-6655aceab8f6-kube-api-access-fzh6b\") pod \"ovn-controller-22dwz\" (UID: \"f4b9eef2-987b-4c54-803c-6655aceab8f6\") " pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.632788 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f4b9eef2-987b-4c54-803c-6655aceab8f6-scripts\") pod \"ovn-controller-22dwz\" (UID: \"f4b9eef2-987b-4c54-803c-6655aceab8f6\") " pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.632922 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f4b9eef2-987b-4c54-803c-6655aceab8f6-var-run-ovn\") pod \"ovn-controller-22dwz\" (UID: \"f4b9eef2-987b-4c54-803c-6655aceab8f6\") " pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.633244 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f4b9eef2-987b-4c54-803c-6655aceab8f6-var-run-ovn\") pod \"ovn-controller-22dwz\" (UID: \"f4b9eef2-987b-4c54-803c-6655aceab8f6\") " pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.633358 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc83d812-aa8d-45e4-b6df-02e6f8a9d9db-config\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.633566 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc83d812-aa8d-45e4-b6df-02e6f8a9d9db-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.633639 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/54ab7f4e-f623-473b-8b6a-3363fe487b4e-etc-ovs\") pod \"ovn-controller-ovs-bdjjv\" (UID: \"54ab7f4e-f623-473b-8b6a-3363fe487b4e\") " pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.633851 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/54ab7f4e-f623-473b-8b6a-3363fe487b4e-etc-ovs\") pod \"ovn-controller-ovs-bdjjv\" (UID: \"54ab7f4e-f623-473b-8b6a-3363fe487b4e\") " pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.633871 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/54ab7f4e-f623-473b-8b6a-3363fe487b4e-var-lib\") pod \"ovn-controller-ovs-bdjjv\" (UID: \"54ab7f4e-f623-473b-8b6a-3363fe487b4e\") " pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.633902 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54ab7f4e-f623-473b-8b6a-3363fe487b4e-scripts\") pod \"ovn-controller-ovs-bdjjv\" (UID: \"54ab7f4e-f623-473b-8b6a-3363fe487b4e\") " pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.634292 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4b9eef2-987b-4c54-803c-6655aceab8f6-ovn-controller-tls-certs\") pod \"ovn-controller-22dwz\" (UID: \"f4b9eef2-987b-4c54-803c-6655aceab8f6\") " pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.634480 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/54ab7f4e-f623-473b-8b6a-3363fe487b4e-var-log\") pod \"ovn-controller-ovs-bdjjv\" (UID: \"54ab7f4e-f623-473b-8b6a-3363fe487b4e\") " pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.634557 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cc83d812-aa8d-45e4-b6df-02e6f8a9d9db-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.634813 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/54ab7f4e-f623-473b-8b6a-3363fe487b4e-var-log\") pod \"ovn-controller-ovs-bdjjv\" (UID: \"54ab7f4e-f623-473b-8b6a-3363fe487b4e\") " pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.634046 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/54ab7f4e-f623-473b-8b6a-3363fe487b4e-var-lib\") pod \"ovn-controller-ovs-bdjjv\" (UID: \"54ab7f4e-f623-473b-8b6a-3363fe487b4e\") " pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.634813 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cc83d812-aa8d-45e4-b6df-02e6f8a9d9db-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.635061 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc83d812-aa8d-45e4-b6df-02e6f8a9d9db-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.635237 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh5cd\" (UniqueName: \"kubernetes.io/projected/cc83d812-aa8d-45e4-b6df-02e6f8a9d9db-kube-api-access-hh5cd\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.635372 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f4b9eef2-987b-4c54-803c-6655aceab8f6-var-log-ovn\") pod \"ovn-controller-22dwz\" (UID: \"f4b9eef2-987b-4c54-803c-6655aceab8f6\") " pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.635389 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b9eef2-987b-4c54-803c-6655aceab8f6-combined-ca-bundle\") pod \"ovn-controller-22dwz\" (UID: \"f4b9eef2-987b-4c54-803c-6655aceab8f6\") " pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.635719 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f4b9eef2-987b-4c54-803c-6655aceab8f6-var-log-ovn\") pod \"ovn-controller-22dwz\" (UID: \"f4b9eef2-987b-4c54-803c-6655aceab8f6\") " pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.635914 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/54ab7f4e-f623-473b-8b6a-3363fe487b4e-scripts\") pod \"ovn-controller-ovs-bdjjv\" (UID: \"54ab7f4e-f623-473b-8b6a-3363fe487b4e\") " pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.642499 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b9eef2-987b-4c54-803c-6655aceab8f6-combined-ca-bundle\") pod \"ovn-controller-22dwz\" (UID: \"f4b9eef2-987b-4c54-803c-6655aceab8f6\") " pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.652875 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4b9eef2-987b-4c54-803c-6655aceab8f6-ovn-controller-tls-certs\") pod \"ovn-controller-22dwz\" (UID: \"f4b9eef2-987b-4c54-803c-6655aceab8f6\") " pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.660191 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m9jp\" (UniqueName: \"kubernetes.io/projected/54ab7f4e-f623-473b-8b6a-3363fe487b4e-kube-api-access-7m9jp\") pod \"ovn-controller-ovs-bdjjv\" (UID: \"54ab7f4e-f623-473b-8b6a-3363fe487b4e\") " pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.664458 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzh6b\" (UniqueName: \"kubernetes.io/projected/f4b9eef2-987b-4c54-803c-6655aceab8f6-kube-api-access-fzh6b\") pod \"ovn-controller-22dwz\" (UID: \"f4b9eef2-987b-4c54-803c-6655aceab8f6\") " pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.728896 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-22dwz" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.736997 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc83d812-aa8d-45e4-b6df-02e6f8a9d9db-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.737033 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh5cd\" (UniqueName: \"kubernetes.io/projected/cc83d812-aa8d-45e4-b6df-02e6f8a9d9db-kube-api-access-hh5cd\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.737069 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc83d812-aa8d-45e4-b6df-02e6f8a9d9db-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.737124 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6c3d800f-2362-434e-a5ce-3e291bfdcf14\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6c3d800f-2362-434e-a5ce-3e291bfdcf14\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.737201 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc83d812-aa8d-45e4-b6df-02e6f8a9d9db-config\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.737220 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc83d812-aa8d-45e4-b6df-02e6f8a9d9db-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.737257 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cc83d812-aa8d-45e4-b6df-02e6f8a9d9db-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.737282 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cc83d812-aa8d-45e4-b6df-02e6f8a9d9db-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.738987 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cc83d812-aa8d-45e4-b6df-02e6f8a9d9db-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.739263 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cc83d812-aa8d-45e4-b6df-02e6f8a9d9db-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.739817 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc83d812-aa8d-45e4-b6df-02e6f8a9d9db-config\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.744333 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc83d812-aa8d-45e4-b6df-02e6f8a9d9db-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.754274 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc83d812-aa8d-45e4-b6df-02e6f8a9d9db-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.754759 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cc83d812-aa8d-45e4-b6df-02e6f8a9d9db-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.754765 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.763594 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh5cd\" (UniqueName: \"kubernetes.io/projected/cc83d812-aa8d-45e4-b6df-02e6f8a9d9db-kube-api-access-hh5cd\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.783803 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.783848 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6c3d800f-2362-434e-a5ce-3e291bfdcf14\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6c3d800f-2362-434e-a5ce-3e291bfdcf14\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1b6a11ad13729865816c2dc5cd4219e6fe66bf31c8930f76476acb4344de05a7/globalmount\"" pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:27 crc kubenswrapper[4934]: I1227 08:02:27.983975 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6c3d800f-2362-434e-a5ce-3e291bfdcf14\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6c3d800f-2362-434e-a5ce-3e291bfdcf14\") pod \"ovsdbserver-nb-0\" (UID: \"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db\") " pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:28 crc kubenswrapper[4934]: I1227 08:02:28.108571 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 27 08:02:30 crc kubenswrapper[4934]: I1227 08:02:30.230460 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 27 08:02:30 crc kubenswrapper[4934]: I1227 08:02:30.788071 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 27 08:02:30 crc kubenswrapper[4934]: I1227 08:02:30.790511 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:30 crc kubenswrapper[4934]: I1227 08:02:30.793815 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 27 08:02:30 crc kubenswrapper[4934]: I1227 08:02:30.794051 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 27 08:02:30 crc kubenswrapper[4934]: I1227 08:02:30.795388 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-f2pqv" Dec 27 08:02:30 crc kubenswrapper[4934]: I1227 08:02:30.795726 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 27 08:02:30 crc kubenswrapper[4934]: I1227 08:02:30.817720 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 27 08:02:30 crc kubenswrapper[4934]: I1227 08:02:30.912830 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/049be8b3-9033-425c-940e-4a36854f06d9-config\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:30 crc kubenswrapper[4934]: I1227 08:02:30.912905 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/049be8b3-9033-425c-940e-4a36854f06d9-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:30 crc kubenswrapper[4934]: I1227 08:02:30.912960 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ab44f442-f5ca-4499-9a0e-e685fc9b5aa1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ab44f442-f5ca-4499-9a0e-e685fc9b5aa1\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:30 crc kubenswrapper[4934]: I1227 08:02:30.912979 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/049be8b3-9033-425c-940e-4a36854f06d9-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:30 crc kubenswrapper[4934]: I1227 08:02:30.912996 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/049be8b3-9033-425c-940e-4a36854f06d9-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:30 crc kubenswrapper[4934]: I1227 08:02:30.913037 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vwfm\" (UniqueName: \"kubernetes.io/projected/049be8b3-9033-425c-940e-4a36854f06d9-kube-api-access-6vwfm\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:30 crc kubenswrapper[4934]: I1227 08:02:30.913138 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/049be8b3-9033-425c-940e-4a36854f06d9-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:30 crc kubenswrapper[4934]: I1227 08:02:30.913164 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/049be8b3-9033-425c-940e-4a36854f06d9-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:31 crc kubenswrapper[4934]: I1227 08:02:31.014877 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/049be8b3-9033-425c-940e-4a36854f06d9-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:31 crc kubenswrapper[4934]: I1227 08:02:31.014941 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/049be8b3-9033-425c-940e-4a36854f06d9-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:31 crc kubenswrapper[4934]: I1227 08:02:31.014974 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/049be8b3-9033-425c-940e-4a36854f06d9-config\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:31 crc kubenswrapper[4934]: I1227 08:02:31.015026 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/049be8b3-9033-425c-940e-4a36854f06d9-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:31 crc kubenswrapper[4934]: I1227 08:02:31.015136 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ab44f442-f5ca-4499-9a0e-e685fc9b5aa1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ab44f442-f5ca-4499-9a0e-e685fc9b5aa1\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:31 crc kubenswrapper[4934]: I1227 08:02:31.015161 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/049be8b3-9033-425c-940e-4a36854f06d9-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:31 crc kubenswrapper[4934]: I1227 08:02:31.015186 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/049be8b3-9033-425c-940e-4a36854f06d9-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:31 crc kubenswrapper[4934]: I1227 08:02:31.015242 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vwfm\" (UniqueName: \"kubernetes.io/projected/049be8b3-9033-425c-940e-4a36854f06d9-kube-api-access-6vwfm\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:31 crc kubenswrapper[4934]: I1227 08:02:31.015479 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/049be8b3-9033-425c-940e-4a36854f06d9-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:31 crc kubenswrapper[4934]: I1227 08:02:31.016022 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/049be8b3-9033-425c-940e-4a36854f06d9-config\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:31 crc kubenswrapper[4934]: I1227 08:02:31.016656 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/049be8b3-9033-425c-940e-4a36854f06d9-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:31 crc kubenswrapper[4934]: I1227 08:02:31.019948 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 08:02:31 crc kubenswrapper[4934]: I1227 08:02:31.020076 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ab44f442-f5ca-4499-9a0e-e685fc9b5aa1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ab44f442-f5ca-4499-9a0e-e685fc9b5aa1\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/53b7f77874dc8290f03249a1acea9d3d8923d8e61a91ec8037a528ebd6b6c2b0/globalmount\"" pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:31 crc kubenswrapper[4934]: I1227 08:02:31.022292 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/049be8b3-9033-425c-940e-4a36854f06d9-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:31 crc kubenswrapper[4934]: I1227 08:02:31.029639 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/049be8b3-9033-425c-940e-4a36854f06d9-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:31 crc kubenswrapper[4934]: I1227 08:02:31.030891 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/049be8b3-9033-425c-940e-4a36854f06d9-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:31 crc kubenswrapper[4934]: I1227 08:02:31.034703 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vwfm\" (UniqueName: \"kubernetes.io/projected/049be8b3-9033-425c-940e-4a36854f06d9-kube-api-access-6vwfm\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:31 crc kubenswrapper[4934]: I1227 08:02:31.088063 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ab44f442-f5ca-4499-9a0e-e685fc9b5aa1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ab44f442-f5ca-4499-9a0e-e685fc9b5aa1\") pod \"ovsdbserver-sb-0\" (UID: \"049be8b3-9033-425c-940e-4a36854f06d9\") " pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:31 crc kubenswrapper[4934]: I1227 08:02:31.116090 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 27 08:02:37 crc kubenswrapper[4934]: I1227 08:02:37.513890 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ca95a4ef-282d-4c26-983f-d7d00f90443c","Type":"ContainerStarted","Data":"4944950391acb898c18c4ea3f2f1529f8edfc5bd498bdda193bf423b4216fc90"} Dec 27 08:02:38 crc kubenswrapper[4934]: E1227 08:02:38.405046 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 27 08:02:38 crc kubenswrapper[4934]: E1227 08:02:38.405329 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5knj2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-kmx4s_openstack(74c5f0ab-04fb-4a95-a88a-8c448309e7fa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:02:38 crc kubenswrapper[4934]: E1227 08:02:38.407180 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-kmx4s" podUID="74c5f0ab-04fb-4a95-a88a-8c448309e7fa" Dec 27 08:02:38 crc kubenswrapper[4934]: E1227 08:02:38.494327 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 27 08:02:38 crc kubenswrapper[4934]: E1227 08:02:38.494902 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cnfgw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-rsz7l_openstack(3397fbf2-3f34-48bf-8c71-aa097ae68716): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:02:38 crc kubenswrapper[4934]: E1227 08:02:38.496224 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-rsz7l" podUID="3397fbf2-3f34-48bf-8c71-aa097ae68716" Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.235107 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.265347 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-kmx4s" Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.318154 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-rsz7l" Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.369802 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5knj2\" (UniqueName: \"kubernetes.io/projected/74c5f0ab-04fb-4a95-a88a-8c448309e7fa-kube-api-access-5knj2\") pod \"74c5f0ab-04fb-4a95-a88a-8c448309e7fa\" (UID: \"74c5f0ab-04fb-4a95-a88a-8c448309e7fa\") " Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.369954 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3397fbf2-3f34-48bf-8c71-aa097ae68716-config\") pod \"3397fbf2-3f34-48bf-8c71-aa097ae68716\" (UID: \"3397fbf2-3f34-48bf-8c71-aa097ae68716\") " Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.370014 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74c5f0ab-04fb-4a95-a88a-8c448309e7fa-dns-svc\") pod \"74c5f0ab-04fb-4a95-a88a-8c448309e7fa\" (UID: \"74c5f0ab-04fb-4a95-a88a-8c448309e7fa\") " Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.370054 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnfgw\" (UniqueName: \"kubernetes.io/projected/3397fbf2-3f34-48bf-8c71-aa097ae68716-kube-api-access-cnfgw\") pod \"3397fbf2-3f34-48bf-8c71-aa097ae68716\" (UID: \"3397fbf2-3f34-48bf-8c71-aa097ae68716\") " Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.370162 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74c5f0ab-04fb-4a95-a88a-8c448309e7fa-config\") pod \"74c5f0ab-04fb-4a95-a88a-8c448309e7fa\" (UID: \"74c5f0ab-04fb-4a95-a88a-8c448309e7fa\") " Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.370806 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74c5f0ab-04fb-4a95-a88a-8c448309e7fa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "74c5f0ab-04fb-4a95-a88a-8c448309e7fa" (UID: "74c5f0ab-04fb-4a95-a88a-8c448309e7fa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.371027 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74c5f0ab-04fb-4a95-a88a-8c448309e7fa-config" (OuterVolumeSpecName: "config") pod "74c5f0ab-04fb-4a95-a88a-8c448309e7fa" (UID: "74c5f0ab-04fb-4a95-a88a-8c448309e7fa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.371443 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3397fbf2-3f34-48bf-8c71-aa097ae68716-config" (OuterVolumeSpecName: "config") pod "3397fbf2-3f34-48bf-8c71-aa097ae68716" (UID: "3397fbf2-3f34-48bf-8c71-aa097ae68716"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.375852 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74c5f0ab-04fb-4a95-a88a-8c448309e7fa-kube-api-access-5knj2" (OuterVolumeSpecName: "kube-api-access-5knj2") pod "74c5f0ab-04fb-4a95-a88a-8c448309e7fa" (UID: "74c5f0ab-04fb-4a95-a88a-8c448309e7fa"). InnerVolumeSpecName "kube-api-access-5knj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.376658 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3397fbf2-3f34-48bf-8c71-aa097ae68716-kube-api-access-cnfgw" (OuterVolumeSpecName: "kube-api-access-cnfgw") pod "3397fbf2-3f34-48bf-8c71-aa097ae68716" (UID: "3397fbf2-3f34-48bf-8c71-aa097ae68716"). InnerVolumeSpecName "kube-api-access-cnfgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.471929 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74c5f0ab-04fb-4a95-a88a-8c448309e7fa-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.471965 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5knj2\" (UniqueName: \"kubernetes.io/projected/74c5f0ab-04fb-4a95-a88a-8c448309e7fa-kube-api-access-5knj2\") on node \"crc\" DevicePath \"\"" Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.471975 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3397fbf2-3f34-48bf-8c71-aa097ae68716-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.471986 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74c5f0ab-04fb-4a95-a88a-8c448309e7fa-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.471994 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnfgw\" (UniqueName: \"kubernetes.io/projected/3397fbf2-3f34-48bf-8c71-aa097ae68716-kube-api-access-cnfgw\") on node \"crc\" DevicePath \"\"" Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.548018 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d03cf582-2192-4a10-b21e-10befebde419","Type":"ContainerStarted","Data":"b8ff4733927f8a9bffb5bd6f27121df57686cdb311917f0f8daa2889c4d24d5a"} Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.552449 4934 generic.go:334] "Generic (PLEG): container finished" podID="ff5d280b-73d2-49a6-bb58-4e515e88e179" containerID="f1eae70f52a625a0edd84551a37f20eb6a0dd013a09893e9af7620aafa533bb1" exitCode=0 Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.552534 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" event={"ID":"ff5d280b-73d2-49a6-bb58-4e515e88e179","Type":"ContainerDied","Data":"f1eae70f52a625a0edd84551a37f20eb6a0dd013a09893e9af7620aafa533bb1"} Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.555141 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ca95a4ef-282d-4c26-983f-d7d00f90443c","Type":"ContainerStarted","Data":"849530c00021e5d36bd1a69eeb472ba14bb87b47832d2d2baf54d09fa204c96a"} Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.557396 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92d66221-653f-4837-a733-f02073b6ed70","Type":"ContainerStarted","Data":"7e838cf66fc939ea1244896217720d1205d8a0cdeb45401e1a154aabbe99ae2c"} Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.558731 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-rsz7l" event={"ID":"3397fbf2-3f34-48bf-8c71-aa097ae68716","Type":"ContainerDied","Data":"2d987ecc491176028f4dfc538d1a7aa995a7aeff2d2b7d04ebce45a0e4a81a7f"} Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.558819 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-rsz7l" Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.560793 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-kmx4s" event={"ID":"74c5f0ab-04fb-4a95-a88a-8c448309e7fa","Type":"ContainerDied","Data":"d3bcafe34da41a92388ef6ceded11435aac1a4133e28fd72c6cdcf04e2f85fd1"} Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.560830 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-kmx4s" Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.565565 4934 generic.go:334] "Generic (PLEG): container finished" podID="ca9c1959-3f6a-4c4d-ab56-b464625d92c8" containerID="6cbabc9006c4fee872f9dd8e09e62ed1b75c67e72fe49d76d2fc6a2033c09151" exitCode=0 Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.565607 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" event={"ID":"ca9c1959-3f6a-4c4d-ab56-b464625d92c8","Type":"ContainerDied","Data":"6cbabc9006c4fee872f9dd8e09e62ed1b75c67e72fe49d76d2fc6a2033c09151"} Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.634899 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-rsz7l"] Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.653325 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-rsz7l"] Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.701611 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-kmx4s"] Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.708988 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-kmx4s"] Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.911707 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.927613 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-74cc7c486-2dpkp"] Dec 27 08:02:39 crc kubenswrapper[4934]: I1227 08:02:39.935210 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-22dwz"] Dec 27 08:02:40 crc kubenswrapper[4934]: I1227 08:02:40.054352 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-66cbf594b5-swd4w"] Dec 27 08:02:40 crc kubenswrapper[4934]: I1227 08:02:40.065183 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 27 08:02:40 crc kubenswrapper[4934]: I1227 08:02:40.097363 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 27 08:02:40 crc kubenswrapper[4934]: W1227 08:02:40.308284 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb102f157_6ec2_4b75_846d_8362a74bf9d2.slice/crio-c3c8db5a39c4b1a791ae41c515a0da37fed02c63b4aac5dd5ae520c79daaf005 WatchSource:0}: Error finding container c3c8db5a39c4b1a791ae41c515a0da37fed02c63b4aac5dd5ae520c79daaf005: Status 404 returned error can't find the container with id c3c8db5a39c4b1a791ae41c515a0da37fed02c63b4aac5dd5ae520c79daaf005 Dec 27 08:02:40 crc kubenswrapper[4934]: I1227 08:02:40.620617 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"049be8b3-9033-425c-940e-4a36854f06d9","Type":"ContainerStarted","Data":"8bbd3f47b04395ac6d1033e5ac84b7c4280a97a44f1b36ad15398497d5f4cc69"} Dec 27 08:02:40 crc kubenswrapper[4934]: I1227 08:02:40.627752 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-66cbf594b5-swd4w" event={"ID":"355bcd5e-7852-477f-8e4b-b03847566cc6","Type":"ContainerStarted","Data":"ec52ecee0420b38855deda058a5783eb3d2c5543bc919ce68a71a29d2898bf66"} Dec 27 08:02:40 crc kubenswrapper[4934]: I1227 08:02:40.646468 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"5788f6c6-2500-4c66-af7a-830cde17caa4","Type":"ContainerStarted","Data":"025c6e94fc16f8dfe3303fd225b5fc2e02f9952414fc805476597fdb089f791b"} Dec 27 08:02:40 crc kubenswrapper[4934]: I1227 08:02:40.648320 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b102f157-6ec2-4b75-846d-8362a74bf9d2","Type":"ContainerStarted","Data":"c3c8db5a39c4b1a791ae41c515a0da37fed02c63b4aac5dd5ae520c79daaf005"} Dec 27 08:02:40 crc kubenswrapper[4934]: I1227 08:02:40.654602 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"26cea6d5-e2df-46ba-a731-722d91bd3c91","Type":"ContainerStarted","Data":"c5e933b18911653b5ac892d1e3730195c01f29c6b25d58fff0297c6dca4d6db3"} Dec 27 08:02:40 crc kubenswrapper[4934]: I1227 08:02:40.669256 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-22dwz" event={"ID":"f4b9eef2-987b-4c54-803c-6655aceab8f6","Type":"ContainerStarted","Data":"1412debaa486c6b7ee0c2e86da2afea46b5338fd4d11c2677c124f7d1ff0c6ca"} Dec 27 08:02:40 crc kubenswrapper[4934]: I1227 08:02:40.692276 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"df7d6b06-cb87-4635-8aca-1ab00a564dd7","Type":"ContainerStarted","Data":"6386108486d5f92741eb22a890f1a957ebb07da8c2f70cbbed654caafe93c1c2"} Dec 27 08:02:40 crc kubenswrapper[4934]: I1227 08:02:40.693509 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-bdjjv"] Dec 27 08:02:40 crc kubenswrapper[4934]: I1227 08:02:40.695860 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74cc7c486-2dpkp" event={"ID":"e0978ad2-1582-49d4-b239-ff929dafb117","Type":"ContainerStarted","Data":"c7ce6dc0ec8b32aef2a6c14dba3ea5a4e4dfbeb8288d04b4d4c4d7d1d9f1a175"} Dec 27 08:02:40 crc kubenswrapper[4934]: I1227 08:02:40.923889 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 27 08:02:41 crc kubenswrapper[4934]: I1227 08:02:41.481846 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3397fbf2-3f34-48bf-8c71-aa097ae68716" path="/var/lib/kubelet/pods/3397fbf2-3f34-48bf-8c71-aa097ae68716/volumes" Dec 27 08:02:41 crc kubenswrapper[4934]: I1227 08:02:41.482369 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74c5f0ab-04fb-4a95-a88a-8c448309e7fa" path="/var/lib/kubelet/pods/74c5f0ab-04fb-4a95-a88a-8c448309e7fa/volumes" Dec 27 08:02:41 crc kubenswrapper[4934]: I1227 08:02:41.705371 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bdjjv" event={"ID":"54ab7f4e-f623-473b-8b6a-3363fe487b4e","Type":"ContainerStarted","Data":"ce10ac48afede45a5c01789bef49b21d6b448225fd5d5e30d7b2099eb7019239"} Dec 27 08:02:41 crc kubenswrapper[4934]: I1227 08:02:41.706571 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db","Type":"ContainerStarted","Data":"c3cc60fc99a7440ac3dd43201b268fd41f2c597d455f976f3ec2d709a8214406"} Dec 27 08:02:42 crc kubenswrapper[4934]: I1227 08:02:42.717524 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74cc7c486-2dpkp" event={"ID":"e0978ad2-1582-49d4-b239-ff929dafb117","Type":"ContainerStarted","Data":"181453bd02c117ae9b8347124cc3e9420a7cb336835f2e3b25945d5e60dafc6c"} Dec 27 08:02:42 crc kubenswrapper[4934]: I1227 08:02:42.721411 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" event={"ID":"ff5d280b-73d2-49a6-bb58-4e515e88e179","Type":"ContainerStarted","Data":"8875dce0d311d4772ec1d4f843b142fdabfe56d537b9eca412402ff18cba10b3"} Dec 27 08:02:42 crc kubenswrapper[4934]: I1227 08:02:42.721898 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" Dec 27 08:02:42 crc kubenswrapper[4934]: I1227 08:02:42.724789 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"76158d63-2099-4418-82a2-09af0432c1a3","Type":"ContainerStarted","Data":"e6252db8825d6b9fe04ede9d278e6c3c2fa0bc3d3cf385da4ee6ce5043dc6b19"} Dec 27 08:02:42 crc kubenswrapper[4934]: I1227 08:02:42.726846 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c","Type":"ContainerStarted","Data":"9e896bff961270fe6bc84ed6b42b91ba71952e4040cd6b4594ed7c4e6be8c873"} Dec 27 08:02:42 crc kubenswrapper[4934]: I1227 08:02:42.732101 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" event={"ID":"ca9c1959-3f6a-4c4d-ab56-b464625d92c8","Type":"ContainerStarted","Data":"8ef7f2c1ecadd3836be5e7f839be16b66848904a88d51eb92ee6d74ee50d8f77"} Dec 27 08:02:42 crc kubenswrapper[4934]: I1227 08:02:42.732461 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" Dec 27 08:02:42 crc kubenswrapper[4934]: I1227 08:02:42.753429 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-74cc7c486-2dpkp" podStartSLOduration=18.753407388 podStartE2EDuration="18.753407388s" podCreationTimestamp="2025-12-27 08:02:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:02:42.740433703 +0000 UTC m=+1223.560874307" watchObservedRunningTime="2025-12-27 08:02:42.753407388 +0000 UTC m=+1223.573848002" Dec 27 08:02:42 crc kubenswrapper[4934]: I1227 08:02:42.802309 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" podStartSLOduration=5.346036656 podStartE2EDuration="25.802284034s" podCreationTimestamp="2025-12-27 08:02:17 +0000 UTC" firstStartedPulling="2025-12-27 08:02:18.268238442 +0000 UTC m=+1199.088679036" lastFinishedPulling="2025-12-27 08:02:38.72448582 +0000 UTC m=+1219.544926414" observedRunningTime="2025-12-27 08:02:42.797778561 +0000 UTC m=+1223.618219155" watchObservedRunningTime="2025-12-27 08:02:42.802284034 +0000 UTC m=+1223.622724628" Dec 27 08:02:42 crc kubenswrapper[4934]: I1227 08:02:42.864925 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" podStartSLOduration=5.17994372 podStartE2EDuration="25.864906765s" podCreationTimestamp="2025-12-27 08:02:17 +0000 UTC" firstStartedPulling="2025-12-27 08:02:17.962512453 +0000 UTC m=+1198.782953047" lastFinishedPulling="2025-12-27 08:02:38.647475498 +0000 UTC m=+1219.467916092" observedRunningTime="2025-12-27 08:02:42.855681194 +0000 UTC m=+1223.676121798" watchObservedRunningTime="2025-12-27 08:02:42.864906765 +0000 UTC m=+1223.685347359" Dec 27 08:02:43 crc kubenswrapper[4934]: I1227 08:02:43.743072 4934 generic.go:334] "Generic (PLEG): container finished" podID="d03cf582-2192-4a10-b21e-10befebde419" containerID="b8ff4733927f8a9bffb5bd6f27121df57686cdb311917f0f8daa2889c4d24d5a" exitCode=0 Dec 27 08:02:43 crc kubenswrapper[4934]: I1227 08:02:43.743123 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d03cf582-2192-4a10-b21e-10befebde419","Type":"ContainerDied","Data":"b8ff4733927f8a9bffb5bd6f27121df57686cdb311917f0f8daa2889c4d24d5a"} Dec 27 08:02:44 crc kubenswrapper[4934]: I1227 08:02:44.758233 4934 generic.go:334] "Generic (PLEG): container finished" podID="ca95a4ef-282d-4c26-983f-d7d00f90443c" containerID="849530c00021e5d36bd1a69eeb472ba14bb87b47832d2d2baf54d09fa204c96a" exitCode=0 Dec 27 08:02:44 crc kubenswrapper[4934]: I1227 08:02:44.758413 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ca95a4ef-282d-4c26-983f-d7d00f90443c","Type":"ContainerDied","Data":"849530c00021e5d36bd1a69eeb472ba14bb87b47832d2d2baf54d09fa204c96a"} Dec 27 08:02:44 crc kubenswrapper[4934]: I1227 08:02:44.985060 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:44 crc kubenswrapper[4934]: I1227 08:02:44.985209 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:44 crc kubenswrapper[4934]: I1227 08:02:44.993403 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:45 crc kubenswrapper[4934]: I1227 08:02:45.771296 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 08:02:45 crc kubenswrapper[4934]: I1227 08:02:45.834252 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-86b5d5c74c-c9lp9"] Dec 27 08:02:47 crc kubenswrapper[4934]: I1227 08:02:47.395874 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" Dec 27 08:02:47 crc kubenswrapper[4934]: I1227 08:02:47.729557 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" Dec 27 08:02:47 crc kubenswrapper[4934]: I1227 08:02:47.842371 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-6lbgv"] Dec 27 08:02:47 crc kubenswrapper[4934]: I1227 08:02:47.842630 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" podUID="ff5d280b-73d2-49a6-bb58-4e515e88e179" containerName="dnsmasq-dns" containerID="cri-o://8875dce0d311d4772ec1d4f843b142fdabfe56d537b9eca412402ff18cba10b3" gracePeriod=10 Dec 27 08:02:48 crc kubenswrapper[4934]: I1227 08:02:48.833945 4934 generic.go:334] "Generic (PLEG): container finished" podID="ff5d280b-73d2-49a6-bb58-4e515e88e179" containerID="8875dce0d311d4772ec1d4f843b142fdabfe56d537b9eca412402ff18cba10b3" exitCode=0 Dec 27 08:02:48 crc kubenswrapper[4934]: I1227 08:02:48.834007 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" event={"ID":"ff5d280b-73d2-49a6-bb58-4e515e88e179","Type":"ContainerDied","Data":"8875dce0d311d4772ec1d4f843b142fdabfe56d537b9eca412402ff18cba10b3"} Dec 27 08:02:49 crc kubenswrapper[4934]: I1227 08:02:49.309905 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" Dec 27 08:02:49 crc kubenswrapper[4934]: I1227 08:02:49.455887 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff5d280b-73d2-49a6-bb58-4e515e88e179-config\") pod \"ff5d280b-73d2-49a6-bb58-4e515e88e179\" (UID: \"ff5d280b-73d2-49a6-bb58-4e515e88e179\") " Dec 27 08:02:49 crc kubenswrapper[4934]: I1227 08:02:49.456070 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hczdl\" (UniqueName: \"kubernetes.io/projected/ff5d280b-73d2-49a6-bb58-4e515e88e179-kube-api-access-hczdl\") pod \"ff5d280b-73d2-49a6-bb58-4e515e88e179\" (UID: \"ff5d280b-73d2-49a6-bb58-4e515e88e179\") " Dec 27 08:02:49 crc kubenswrapper[4934]: I1227 08:02:49.456171 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff5d280b-73d2-49a6-bb58-4e515e88e179-dns-svc\") pod \"ff5d280b-73d2-49a6-bb58-4e515e88e179\" (UID: \"ff5d280b-73d2-49a6-bb58-4e515e88e179\") " Dec 27 08:02:49 crc kubenswrapper[4934]: I1227 08:02:49.567226 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff5d280b-73d2-49a6-bb58-4e515e88e179-kube-api-access-hczdl" (OuterVolumeSpecName: "kube-api-access-hczdl") pod "ff5d280b-73d2-49a6-bb58-4e515e88e179" (UID: "ff5d280b-73d2-49a6-bb58-4e515e88e179"). InnerVolumeSpecName "kube-api-access-hczdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:02:49 crc kubenswrapper[4934]: I1227 08:02:49.660505 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hczdl\" (UniqueName: \"kubernetes.io/projected/ff5d280b-73d2-49a6-bb58-4e515e88e179-kube-api-access-hczdl\") on node \"crc\" DevicePath \"\"" Dec 27 08:02:49 crc kubenswrapper[4934]: I1227 08:02:49.846274 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" event={"ID":"ff5d280b-73d2-49a6-bb58-4e515e88e179","Type":"ContainerDied","Data":"e3ec4fc7e198727a2dd293607963b9e6f3a6f60a4a28ddec6396d7d2a06229c9"} Dec 27 08:02:49 crc kubenswrapper[4934]: I1227 08:02:49.846341 4934 scope.go:117] "RemoveContainer" containerID="8875dce0d311d4772ec1d4f843b142fdabfe56d537b9eca412402ff18cba10b3" Dec 27 08:02:49 crc kubenswrapper[4934]: I1227 08:02:49.846358 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-6lbgv" Dec 27 08:02:49 crc kubenswrapper[4934]: I1227 08:02:49.969722 4934 scope.go:117] "RemoveContainer" containerID="f1eae70f52a625a0edd84551a37f20eb6a0dd013a09893e9af7620aafa533bb1" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.152305 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff5d280b-73d2-49a6-bb58-4e515e88e179-config" (OuterVolumeSpecName: "config") pod "ff5d280b-73d2-49a6-bb58-4e515e88e179" (UID: "ff5d280b-73d2-49a6-bb58-4e515e88e179"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.176201 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff5d280b-73d2-49a6-bb58-4e515e88e179-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.201000 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff5d280b-73d2-49a6-bb58-4e515e88e179-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ff5d280b-73d2-49a6-bb58-4e515e88e179" (UID: "ff5d280b-73d2-49a6-bb58-4e515e88e179"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.278499 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff5d280b-73d2-49a6-bb58-4e515e88e179-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.478021 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-6lbgv"] Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.487151 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-6lbgv"] Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.546420 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-5v4n5"] Dec 27 08:02:50 crc kubenswrapper[4934]: E1227 08:02:50.547153 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff5d280b-73d2-49a6-bb58-4e515e88e179" containerName="init" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.547248 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff5d280b-73d2-49a6-bb58-4e515e88e179" containerName="init" Dec 27 08:02:50 crc kubenswrapper[4934]: E1227 08:02:50.547318 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff5d280b-73d2-49a6-bb58-4e515e88e179" containerName="dnsmasq-dns" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.547379 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff5d280b-73d2-49a6-bb58-4e515e88e179" containerName="dnsmasq-dns" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.547600 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff5d280b-73d2-49a6-bb58-4e515e88e179" containerName="dnsmasq-dns" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.548297 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-5v4n5" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.550868 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.588136 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-5v4n5"] Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.686226 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/36e93e22-f392-4011-a82a-a5d1c26b5970-ovs-rundir\") pod \"ovn-controller-metrics-5v4n5\" (UID: \"36e93e22-f392-4011-a82a-a5d1c26b5970\") " pod="openstack/ovn-controller-metrics-5v4n5" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.686975 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/36e93e22-f392-4011-a82a-a5d1c26b5970-ovn-rundir\") pod \"ovn-controller-metrics-5v4n5\" (UID: \"36e93e22-f392-4011-a82a-a5d1c26b5970\") " pod="openstack/ovn-controller-metrics-5v4n5" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.687060 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/36e93e22-f392-4011-a82a-a5d1c26b5970-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-5v4n5\" (UID: \"36e93e22-f392-4011-a82a-a5d1c26b5970\") " pod="openstack/ovn-controller-metrics-5v4n5" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.687223 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36e93e22-f392-4011-a82a-a5d1c26b5970-config\") pod \"ovn-controller-metrics-5v4n5\" (UID: \"36e93e22-f392-4011-a82a-a5d1c26b5970\") " pod="openstack/ovn-controller-metrics-5v4n5" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.687382 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36e93e22-f392-4011-a82a-a5d1c26b5970-combined-ca-bundle\") pod \"ovn-controller-metrics-5v4n5\" (UID: \"36e93e22-f392-4011-a82a-a5d1c26b5970\") " pod="openstack/ovn-controller-metrics-5v4n5" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.687576 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g94l7\" (UniqueName: \"kubernetes.io/projected/36e93e22-f392-4011-a82a-a5d1c26b5970-kube-api-access-g94l7\") pod \"ovn-controller-metrics-5v4n5\" (UID: \"36e93e22-f392-4011-a82a-a5d1c26b5970\") " pod="openstack/ovn-controller-metrics-5v4n5" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.692186 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-5tspd"] Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.693958 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-5tspd" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.696388 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.717155 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-5tspd"] Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.789198 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-5tspd\" (UID: \"cbe083f3-5a60-4b98-b13f-65b82fcae1c1\") " pod="openstack/dnsmasq-dns-6bc7876d45-5tspd" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.789484 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-config\") pod \"dnsmasq-dns-6bc7876d45-5tspd\" (UID: \"cbe083f3-5a60-4b98-b13f-65b82fcae1c1\") " pod="openstack/dnsmasq-dns-6bc7876d45-5tspd" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.789587 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8cd8\" (UniqueName: \"kubernetes.io/projected/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-kube-api-access-z8cd8\") pod \"dnsmasq-dns-6bc7876d45-5tspd\" (UID: \"cbe083f3-5a60-4b98-b13f-65b82fcae1c1\") " pod="openstack/dnsmasq-dns-6bc7876d45-5tspd" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.789712 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/36e93e22-f392-4011-a82a-a5d1c26b5970-ovn-rundir\") pod \"ovn-controller-metrics-5v4n5\" (UID: \"36e93e22-f392-4011-a82a-a5d1c26b5970\") " pod="openstack/ovn-controller-metrics-5v4n5" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.789792 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/36e93e22-f392-4011-a82a-a5d1c26b5970-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-5v4n5\" (UID: \"36e93e22-f392-4011-a82a-a5d1c26b5970\") " pod="openstack/ovn-controller-metrics-5v4n5" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.789893 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36e93e22-f392-4011-a82a-a5d1c26b5970-config\") pod \"ovn-controller-metrics-5v4n5\" (UID: \"36e93e22-f392-4011-a82a-a5d1c26b5970\") " pod="openstack/ovn-controller-metrics-5v4n5" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.790115 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36e93e22-f392-4011-a82a-a5d1c26b5970-combined-ca-bundle\") pod \"ovn-controller-metrics-5v4n5\" (UID: \"36e93e22-f392-4011-a82a-a5d1c26b5970\") " pod="openstack/ovn-controller-metrics-5v4n5" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.790798 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-5tspd\" (UID: \"cbe083f3-5a60-4b98-b13f-65b82fcae1c1\") " pod="openstack/dnsmasq-dns-6bc7876d45-5tspd" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.790897 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g94l7\" (UniqueName: \"kubernetes.io/projected/36e93e22-f392-4011-a82a-a5d1c26b5970-kube-api-access-g94l7\") pod \"ovn-controller-metrics-5v4n5\" (UID: \"36e93e22-f392-4011-a82a-a5d1c26b5970\") " pod="openstack/ovn-controller-metrics-5v4n5" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.790994 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/36e93e22-f392-4011-a82a-a5d1c26b5970-ovs-rundir\") pod \"ovn-controller-metrics-5v4n5\" (UID: \"36e93e22-f392-4011-a82a-a5d1c26b5970\") " pod="openstack/ovn-controller-metrics-5v4n5" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.791189 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/36e93e22-f392-4011-a82a-a5d1c26b5970-ovs-rundir\") pod \"ovn-controller-metrics-5v4n5\" (UID: \"36e93e22-f392-4011-a82a-a5d1c26b5970\") " pod="openstack/ovn-controller-metrics-5v4n5" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.790006 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/36e93e22-f392-4011-a82a-a5d1c26b5970-ovn-rundir\") pod \"ovn-controller-metrics-5v4n5\" (UID: \"36e93e22-f392-4011-a82a-a5d1c26b5970\") " pod="openstack/ovn-controller-metrics-5v4n5" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.790766 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36e93e22-f392-4011-a82a-a5d1c26b5970-config\") pod \"ovn-controller-metrics-5v4n5\" (UID: \"36e93e22-f392-4011-a82a-a5d1c26b5970\") " pod="openstack/ovn-controller-metrics-5v4n5" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.795515 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/36e93e22-f392-4011-a82a-a5d1c26b5970-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-5v4n5\" (UID: \"36e93e22-f392-4011-a82a-a5d1c26b5970\") " pod="openstack/ovn-controller-metrics-5v4n5" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.800951 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36e93e22-f392-4011-a82a-a5d1c26b5970-combined-ca-bundle\") pod \"ovn-controller-metrics-5v4n5\" (UID: \"36e93e22-f392-4011-a82a-a5d1c26b5970\") " pod="openstack/ovn-controller-metrics-5v4n5" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.811622 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g94l7\" (UniqueName: \"kubernetes.io/projected/36e93e22-f392-4011-a82a-a5d1c26b5970-kube-api-access-g94l7\") pod \"ovn-controller-metrics-5v4n5\" (UID: \"36e93e22-f392-4011-a82a-a5d1c26b5970\") " pod="openstack/ovn-controller-metrics-5v4n5" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.859510 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-66cbf594b5-swd4w" event={"ID":"355bcd5e-7852-477f-8e4b-b03847566cc6","Type":"ContainerStarted","Data":"1ff57e5bad2ae8f7b892b092f4e28a4fd11fc1f11086766b3c819ad9359d0df0"} Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.861691 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"26cea6d5-e2df-46ba-a731-722d91bd3c91","Type":"ContainerStarted","Data":"ccfc511f4ea276acde9189dd054526355385fbd2bcc5b07b341eaf9ddbb666ba"} Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.861937 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.864198 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-22dwz" event={"ID":"f4b9eef2-987b-4c54-803c-6655aceab8f6","Type":"ContainerStarted","Data":"4c077448476e7e3d7b9f1ec708a29834733a34a065bf155fa84e015a864cb1b1"} Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.864311 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-22dwz" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.865983 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db","Type":"ContainerStarted","Data":"2801a7bf86d458be55000ca43c567e705146f848873691a8a60d232001773a7e"} Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.866713 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-5v4n5" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.872632 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d03cf582-2192-4a10-b21e-10befebde419","Type":"ContainerStarted","Data":"a04418c8ab5bb1fa91f94e0931bec31f25b20782638b4a478cac208f02a4620d"} Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.884428 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-ui-dashboards-66cbf594b5-swd4w" podStartSLOduration=18.067090714 podStartE2EDuration="26.88440411s" podCreationTimestamp="2025-12-27 08:02:24 +0000 UTC" firstStartedPulling="2025-12-27 08:02:40.256387065 +0000 UTC m=+1221.076827659" lastFinishedPulling="2025-12-27 08:02:49.073700461 +0000 UTC m=+1229.894141055" observedRunningTime="2025-12-27 08:02:50.874327127 +0000 UTC m=+1231.694767721" watchObservedRunningTime="2025-12-27 08:02:50.88440411 +0000 UTC m=+1231.704844704" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.890955 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"049be8b3-9033-425c-940e-4a36854f06d9","Type":"ContainerStarted","Data":"847ec03277f5745947319cb66b66ff3aa6656e0ea772adab88ec11a18d1853fa"} Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.892536 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-5tspd\" (UID: \"cbe083f3-5a60-4b98-b13f-65b82fcae1c1\") " pod="openstack/dnsmasq-dns-6bc7876d45-5tspd" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.892623 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-5tspd\" (UID: \"cbe083f3-5a60-4b98-b13f-65b82fcae1c1\") " pod="openstack/dnsmasq-dns-6bc7876d45-5tspd" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.892646 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-config\") pod \"dnsmasq-dns-6bc7876d45-5tspd\" (UID: \"cbe083f3-5a60-4b98-b13f-65b82fcae1c1\") " pod="openstack/dnsmasq-dns-6bc7876d45-5tspd" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.892668 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8cd8\" (UniqueName: \"kubernetes.io/projected/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-kube-api-access-z8cd8\") pod \"dnsmasq-dns-6bc7876d45-5tspd\" (UID: \"cbe083f3-5a60-4b98-b13f-65b82fcae1c1\") " pod="openstack/dnsmasq-dns-6bc7876d45-5tspd" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.893618 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-5tspd\" (UID: \"cbe083f3-5a60-4b98-b13f-65b82fcae1c1\") " pod="openstack/dnsmasq-dns-6bc7876d45-5tspd" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.894270 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-5tspd\" (UID: \"cbe083f3-5a60-4b98-b13f-65b82fcae1c1\") " pod="openstack/dnsmasq-dns-6bc7876d45-5tspd" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.896619 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-config\") pod \"dnsmasq-dns-6bc7876d45-5tspd\" (UID: \"cbe083f3-5a60-4b98-b13f-65b82fcae1c1\") " pod="openstack/dnsmasq-dns-6bc7876d45-5tspd" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.898917 4934 generic.go:334] "Generic (PLEG): container finished" podID="54ab7f4e-f623-473b-8b6a-3363fe487b4e" containerID="1269925295c9803cf5b0b6efc23c1600ef11e308b1ecc8982469e386ab2a79c9" exitCode=0 Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.899113 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bdjjv" event={"ID":"54ab7f4e-f623-473b-8b6a-3363fe487b4e","Type":"ContainerDied","Data":"1269925295c9803cf5b0b6efc23c1600ef11e308b1ecc8982469e386ab2a79c9"} Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.902358 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ca95a4ef-282d-4c26-983f-d7d00f90443c","Type":"ContainerStarted","Data":"e862df5ec5f94dc038eceb2b1533dcf21f0a38c95f61e193874266056fd27165"} Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.912653 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"b102f157-6ec2-4b75-846d-8362a74bf9d2","Type":"ContainerStarted","Data":"600d1f7f3278de420b04a24b2a729ee8dc79be3a0c33915224bbb7203040c8ce"} Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.913020 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.920384 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=18.831416093 podStartE2EDuration="27.920369092s" podCreationTimestamp="2025-12-27 08:02:23 +0000 UTC" firstStartedPulling="2025-12-27 08:02:40.245633906 +0000 UTC m=+1221.066074500" lastFinishedPulling="2025-12-27 08:02:49.334586905 +0000 UTC m=+1230.155027499" observedRunningTime="2025-12-27 08:02:50.890682617 +0000 UTC m=+1231.711123211" watchObservedRunningTime="2025-12-27 08:02:50.920369092 +0000 UTC m=+1231.740809676" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.928705 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8cd8\" (UniqueName: \"kubernetes.io/projected/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-kube-api-access-z8cd8\") pod \"dnsmasq-dns-6bc7876d45-5tspd\" (UID: \"cbe083f3-5a60-4b98-b13f-65b82fcae1c1\") " pod="openstack/dnsmasq-dns-6bc7876d45-5tspd" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.957782 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=14.84963008 podStartE2EDuration="32.9577637s" podCreationTimestamp="2025-12-27 08:02:18 +0000 UTC" firstStartedPulling="2025-12-27 08:02:20.524101436 +0000 UTC m=+1201.344542030" lastFinishedPulling="2025-12-27 08:02:38.632235056 +0000 UTC m=+1219.452675650" observedRunningTime="2025-12-27 08:02:50.951124793 +0000 UTC m=+1231.771565387" watchObservedRunningTime="2025-12-27 08:02:50.9577637 +0000 UTC m=+1231.778204294" Dec 27 08:02:50 crc kubenswrapper[4934]: I1227 08:02:50.985608 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-22dwz" podStartSLOduration=15.156835795 podStartE2EDuration="23.985577578s" podCreationTimestamp="2025-12-27 08:02:27 +0000 UTC" firstStartedPulling="2025-12-27 08:02:40.255549454 +0000 UTC m=+1221.075990048" lastFinishedPulling="2025-12-27 08:02:49.084291237 +0000 UTC m=+1229.904731831" observedRunningTime="2025-12-27 08:02:50.983707551 +0000 UTC m=+1231.804148145" watchObservedRunningTime="2025-12-27 08:02:50.985577578 +0000 UTC m=+1231.806018172" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.020313 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-5tspd" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.021230 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=29.672164463 podStartE2EDuration="31.021210151s" podCreationTimestamp="2025-12-27 08:02:20 +0000 UTC" firstStartedPulling="2025-12-27 08:02:37.37019086 +0000 UTC m=+1218.190631494" lastFinishedPulling="2025-12-27 08:02:38.719236588 +0000 UTC m=+1219.539677182" observedRunningTime="2025-12-27 08:02:51.018599106 +0000 UTC m=+1231.839039710" watchObservedRunningTime="2025-12-27 08:02:51.021210151 +0000 UTC m=+1231.841650745" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.102728 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=22.653436285 podStartE2EDuration="30.102705406s" podCreationTimestamp="2025-12-27 08:02:21 +0000 UTC" firstStartedPulling="2025-12-27 08:02:40.323773816 +0000 UTC m=+1221.144214410" lastFinishedPulling="2025-12-27 08:02:47.773042937 +0000 UTC m=+1228.593483531" observedRunningTime="2025-12-27 08:02:51.057875391 +0000 UTC m=+1231.878316015" watchObservedRunningTime="2025-12-27 08:02:51.102705406 +0000 UTC m=+1231.923146000" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.192340 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-5tspd"] Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.249236 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-xp7tf"] Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.250861 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-xp7tf" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.253604 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.277573 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-xp7tf"] Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.313705 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-config\") pod \"dnsmasq-dns-8554648995-xp7tf\" (UID: \"c9266772-ff58-4a27-ad46-99f8b6e344bd\") " pod="openstack/dnsmasq-dns-8554648995-xp7tf" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.314012 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-dns-svc\") pod \"dnsmasq-dns-8554648995-xp7tf\" (UID: \"c9266772-ff58-4a27-ad46-99f8b6e344bd\") " pod="openstack/dnsmasq-dns-8554648995-xp7tf" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.314184 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmfh6\" (UniqueName: \"kubernetes.io/projected/c9266772-ff58-4a27-ad46-99f8b6e344bd-kube-api-access-xmfh6\") pod \"dnsmasq-dns-8554648995-xp7tf\" (UID: \"c9266772-ff58-4a27-ad46-99f8b6e344bd\") " pod="openstack/dnsmasq-dns-8554648995-xp7tf" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.314216 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-xp7tf\" (UID: \"c9266772-ff58-4a27-ad46-99f8b6e344bd\") " pod="openstack/dnsmasq-dns-8554648995-xp7tf" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.314255 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-xp7tf\" (UID: \"c9266772-ff58-4a27-ad46-99f8b6e344bd\") " pod="openstack/dnsmasq-dns-8554648995-xp7tf" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.371862 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.372128 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.415572 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-xp7tf\" (UID: \"c9266772-ff58-4a27-ad46-99f8b6e344bd\") " pod="openstack/dnsmasq-dns-8554648995-xp7tf" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.415697 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-config\") pod \"dnsmasq-dns-8554648995-xp7tf\" (UID: \"c9266772-ff58-4a27-ad46-99f8b6e344bd\") " pod="openstack/dnsmasq-dns-8554648995-xp7tf" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.415741 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-dns-svc\") pod \"dnsmasq-dns-8554648995-xp7tf\" (UID: \"c9266772-ff58-4a27-ad46-99f8b6e344bd\") " pod="openstack/dnsmasq-dns-8554648995-xp7tf" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.415807 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmfh6\" (UniqueName: \"kubernetes.io/projected/c9266772-ff58-4a27-ad46-99f8b6e344bd-kube-api-access-xmfh6\") pod \"dnsmasq-dns-8554648995-xp7tf\" (UID: \"c9266772-ff58-4a27-ad46-99f8b6e344bd\") " pod="openstack/dnsmasq-dns-8554648995-xp7tf" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.415832 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-xp7tf\" (UID: \"c9266772-ff58-4a27-ad46-99f8b6e344bd\") " pod="openstack/dnsmasq-dns-8554648995-xp7tf" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.416946 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-xp7tf\" (UID: \"c9266772-ff58-4a27-ad46-99f8b6e344bd\") " pod="openstack/dnsmasq-dns-8554648995-xp7tf" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.417035 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-dns-svc\") pod \"dnsmasq-dns-8554648995-xp7tf\" (UID: \"c9266772-ff58-4a27-ad46-99f8b6e344bd\") " pod="openstack/dnsmasq-dns-8554648995-xp7tf" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.418097 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-xp7tf\" (UID: \"c9266772-ff58-4a27-ad46-99f8b6e344bd\") " pod="openstack/dnsmasq-dns-8554648995-xp7tf" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.418750 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-config\") pod \"dnsmasq-dns-8554648995-xp7tf\" (UID: \"c9266772-ff58-4a27-ad46-99f8b6e344bd\") " pod="openstack/dnsmasq-dns-8554648995-xp7tf" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.447106 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmfh6\" (UniqueName: \"kubernetes.io/projected/c9266772-ff58-4a27-ad46-99f8b6e344bd-kube-api-access-xmfh6\") pod \"dnsmasq-dns-8554648995-xp7tf\" (UID: \"c9266772-ff58-4a27-ad46-99f8b6e344bd\") " pod="openstack/dnsmasq-dns-8554648995-xp7tf" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.499053 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff5d280b-73d2-49a6-bb58-4e515e88e179" path="/var/lib/kubelet/pods/ff5d280b-73d2-49a6-bb58-4e515e88e179/volumes" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.710144 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-xp7tf" Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.863281 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-5tspd"] Dec 27 08:02:51 crc kubenswrapper[4934]: I1227 08:02:51.928259 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-5tspd" event={"ID":"cbe083f3-5a60-4b98-b13f-65b82fcae1c1","Type":"ContainerStarted","Data":"1792e8c0984aec827327ad067d23d10d4ebb0056807487f78e0e175f80a9f1c7"} Dec 27 08:02:52 crc kubenswrapper[4934]: E1227 08:02:52.007318 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54ab7f4e_f623_473b_8b6a_3363fe487b4e.slice/crio-conmon-1269925295c9803cf5b0b6efc23c1600ef11e308b1ecc8982469e386ab2a79c9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54ab7f4e_f623_473b_8b6a_3363fe487b4e.slice/crio-1269925295c9803cf5b0b6efc23c1600ef11e308b1ecc8982469e386ab2a79c9.scope\": RecentStats: unable to find data in memory cache]" Dec 27 08:02:52 crc kubenswrapper[4934]: I1227 08:02:52.016134 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-5v4n5"] Dec 27 08:02:52 crc kubenswrapper[4934]: W1227 08:02:52.358417 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36e93e22_f392_4011_a82a_a5d1c26b5970.slice/crio-9ee073f4c18fab459008f5bfb4bf4d08ad9808c65ddf3e04ee6efffe6ced86a3 WatchSource:0}: Error finding container 9ee073f4c18fab459008f5bfb4bf4d08ad9808c65ddf3e04ee6efffe6ced86a3: Status 404 returned error can't find the container with id 9ee073f4c18fab459008f5bfb4bf4d08ad9808c65ddf3e04ee6efffe6ced86a3 Dec 27 08:02:52 crc kubenswrapper[4934]: I1227 08:02:52.410478 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-xp7tf"] Dec 27 08:02:52 crc kubenswrapper[4934]: I1227 08:02:52.937208 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-xp7tf" event={"ID":"c9266772-ff58-4a27-ad46-99f8b6e344bd","Type":"ContainerStarted","Data":"2c0d3518c5e7906ff6664670cdf309af08307669e6446c4b4f9bf0318c9e9bc6"} Dec 27 08:02:52 crc kubenswrapper[4934]: I1227 08:02:52.938783 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-5v4n5" event={"ID":"36e93e22-f392-4011-a82a-a5d1c26b5970","Type":"ContainerStarted","Data":"9ee073f4c18fab459008f5bfb4bf4d08ad9808c65ddf3e04ee6efffe6ced86a3"} Dec 27 08:02:55 crc kubenswrapper[4934]: I1227 08:02:55.975070 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92d66221-653f-4837-a733-f02073b6ed70","Type":"ContainerStarted","Data":"6984265a9dcba06e22bdc66f68ad44c0acc9f56b5319d26f0929c8d549b57ab3"} Dec 27 08:02:55 crc kubenswrapper[4934]: I1227 08:02:55.981852 4934 generic.go:334] "Generic (PLEG): container finished" podID="c9266772-ff58-4a27-ad46-99f8b6e344bd" containerID="c4ef3840320e11fd8cd6f5eae183bb52ed263737f8d8dfb44351ea6b255802a8" exitCode=0 Dec 27 08:02:55 crc kubenswrapper[4934]: I1227 08:02:55.981936 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-xp7tf" event={"ID":"c9266772-ff58-4a27-ad46-99f8b6e344bd","Type":"ContainerDied","Data":"c4ef3840320e11fd8cd6f5eae183bb52ed263737f8d8dfb44351ea6b255802a8"} Dec 27 08:02:55 crc kubenswrapper[4934]: I1227 08:02:55.985890 4934 generic.go:334] "Generic (PLEG): container finished" podID="cbe083f3-5a60-4b98-b13f-65b82fcae1c1" containerID="bf1d638c847bf60f0aae6b8d3cbb4959f3832761007631b859c0525327faa069" exitCode=0 Dec 27 08:02:55 crc kubenswrapper[4934]: I1227 08:02:55.985930 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-5tspd" event={"ID":"cbe083f3-5a60-4b98-b13f-65b82fcae1c1","Type":"ContainerDied","Data":"bf1d638c847bf60f0aae6b8d3cbb4959f3832761007631b859c0525327faa069"} Dec 27 08:02:55 crc kubenswrapper[4934]: I1227 08:02:55.989762 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bdjjv" event={"ID":"54ab7f4e-f623-473b-8b6a-3363fe487b4e","Type":"ContainerStarted","Data":"ab42e8c4fdffc4e91072f5d4b8ee5fdc2c7624d33385d9c211a56289579abc04"} Dec 27 08:02:56 crc kubenswrapper[4934]: I1227 08:02:56.769468 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 27 08:02:56 crc kubenswrapper[4934]: I1227 08:02:56.912451 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-5tspd" Dec 27 08:02:56 crc kubenswrapper[4934]: I1227 08:02:56.971193 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-config\") pod \"cbe083f3-5a60-4b98-b13f-65b82fcae1c1\" (UID: \"cbe083f3-5a60-4b98-b13f-65b82fcae1c1\") " Dec 27 08:02:56 crc kubenswrapper[4934]: I1227 08:02:56.971280 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-ovsdbserver-sb\") pod \"cbe083f3-5a60-4b98-b13f-65b82fcae1c1\" (UID: \"cbe083f3-5a60-4b98-b13f-65b82fcae1c1\") " Dec 27 08:02:56 crc kubenswrapper[4934]: I1227 08:02:56.971306 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-dns-svc\") pod \"cbe083f3-5a60-4b98-b13f-65b82fcae1c1\" (UID: \"cbe083f3-5a60-4b98-b13f-65b82fcae1c1\") " Dec 27 08:02:56 crc kubenswrapper[4934]: I1227 08:02:56.971431 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8cd8\" (UniqueName: \"kubernetes.io/projected/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-kube-api-access-z8cd8\") pod \"cbe083f3-5a60-4b98-b13f-65b82fcae1c1\" (UID: \"cbe083f3-5a60-4b98-b13f-65b82fcae1c1\") " Dec 27 08:02:56 crc kubenswrapper[4934]: I1227 08:02:56.976583 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-kube-api-access-z8cd8" (OuterVolumeSpecName: "kube-api-access-z8cd8") pod "cbe083f3-5a60-4b98-b13f-65b82fcae1c1" (UID: "cbe083f3-5a60-4b98-b13f-65b82fcae1c1"). InnerVolumeSpecName "kube-api-access-z8cd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:02:56 crc kubenswrapper[4934]: I1227 08:02:56.993785 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-config" (OuterVolumeSpecName: "config") pod "cbe083f3-5a60-4b98-b13f-65b82fcae1c1" (UID: "cbe083f3-5a60-4b98-b13f-65b82fcae1c1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:02:56 crc kubenswrapper[4934]: I1227 08:02:56.998460 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cbe083f3-5a60-4b98-b13f-65b82fcae1c1" (UID: "cbe083f3-5a60-4b98-b13f-65b82fcae1c1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:02:57 crc kubenswrapper[4934]: I1227 08:02:57.003384 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cbe083f3-5a60-4b98-b13f-65b82fcae1c1" (UID: "cbe083f3-5a60-4b98-b13f-65b82fcae1c1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:02:57 crc kubenswrapper[4934]: I1227 08:02:57.008213 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-5tspd" Dec 27 08:02:57 crc kubenswrapper[4934]: I1227 08:02:57.008317 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-5tspd" event={"ID":"cbe083f3-5a60-4b98-b13f-65b82fcae1c1","Type":"ContainerDied","Data":"1792e8c0984aec827327ad067d23d10d4ebb0056807487f78e0e175f80a9f1c7"} Dec 27 08:02:57 crc kubenswrapper[4934]: I1227 08:02:57.008354 4934 scope.go:117] "RemoveContainer" containerID="bf1d638c847bf60f0aae6b8d3cbb4959f3832761007631b859c0525327faa069" Dec 27 08:02:57 crc kubenswrapper[4934]: I1227 08:02:57.073784 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8cd8\" (UniqueName: \"kubernetes.io/projected/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-kube-api-access-z8cd8\") on node \"crc\" DevicePath \"\"" Dec 27 08:02:57 crc kubenswrapper[4934]: I1227 08:02:57.073809 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:02:57 crc kubenswrapper[4934]: I1227 08:02:57.073819 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 27 08:02:57 crc kubenswrapper[4934]: I1227 08:02:57.073828 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbe083f3-5a60-4b98-b13f-65b82fcae1c1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 27 08:02:57 crc kubenswrapper[4934]: I1227 08:02:57.112953 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-5tspd"] Dec 27 08:02:57 crc kubenswrapper[4934]: I1227 08:02:57.125394 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-5tspd"] Dec 27 08:02:57 crc kubenswrapper[4934]: I1227 08:02:57.484422 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbe083f3-5a60-4b98-b13f-65b82fcae1c1" path="/var/lib/kubelet/pods/cbe083f3-5a60-4b98-b13f-65b82fcae1c1/volumes" Dec 27 08:02:58 crc kubenswrapper[4934]: I1227 08:02:58.031647 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bdjjv" event={"ID":"54ab7f4e-f623-473b-8b6a-3363fe487b4e","Type":"ContainerStarted","Data":"0c3fdd7d884a2803cc59ec0673ce924f7c8b37daf08cf1fc5efc5430b66c820f"} Dec 27 08:02:58 crc kubenswrapper[4934]: I1227 08:02:58.032467 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:58 crc kubenswrapper[4934]: I1227 08:02:58.053344 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-bdjjv" podStartSLOduration=23.329279178 podStartE2EDuration="31.05332321s" podCreationTimestamp="2025-12-27 08:02:27 +0000 UTC" firstStartedPulling="2025-12-27 08:02:40.701275725 +0000 UTC m=+1221.521716319" lastFinishedPulling="2025-12-27 08:02:48.425319747 +0000 UTC m=+1229.245760351" observedRunningTime="2025-12-27 08:02:58.048197501 +0000 UTC m=+1238.868638115" watchObservedRunningTime="2025-12-27 08:02:58.05332321 +0000 UTC m=+1238.873763804" Dec 27 08:02:59 crc kubenswrapper[4934]: I1227 08:02:59.043972 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:02:59 crc kubenswrapper[4934]: I1227 08:02:59.622629 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 27 08:02:59 crc kubenswrapper[4934]: I1227 08:02:59.686304 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 27 08:03:00 crc kubenswrapper[4934]: I1227 08:03:00.031273 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 27 08:03:00 crc kubenswrapper[4934]: I1227 08:03:00.031316 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 27 08:03:00 crc kubenswrapper[4934]: I1227 08:03:00.057667 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"049be8b3-9033-425c-940e-4a36854f06d9","Type":"ContainerStarted","Data":"1ce046eb25fe610cae51225f20682b1d70de776ec505dc1fd4fcbaafe215d3b1"} Dec 27 08:03:00 crc kubenswrapper[4934]: I1227 08:03:00.059590 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-5v4n5" event={"ID":"36e93e22-f392-4011-a82a-a5d1c26b5970","Type":"ContainerStarted","Data":"35207707dbbdbfe54093215bbd3a6901ff6ba90e7307e3d8c5f246767cb024b1"} Dec 27 08:03:00 crc kubenswrapper[4934]: I1227 08:03:00.061613 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-xp7tf" event={"ID":"c9266772-ff58-4a27-ad46-99f8b6e344bd","Type":"ContainerStarted","Data":"ca21fca8f6f9ffe20c122ba46c188e52c5a204d1c7b7d64b8d8308ee9eaad95b"} Dec 27 08:03:00 crc kubenswrapper[4934]: I1227 08:03:00.062460 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-xp7tf" Dec 27 08:03:00 crc kubenswrapper[4934]: I1227 08:03:00.063733 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"cc83d812-aa8d-45e4-b6df-02e6f8a9d9db","Type":"ContainerStarted","Data":"04e05d6c1a818e27e157c42a0fa32b8bf5130a825118bc491df9deaa75b82799"} Dec 27 08:03:00 crc kubenswrapper[4934]: I1227 08:03:00.079757 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=13.59199937 podStartE2EDuration="31.079732789s" podCreationTimestamp="2025-12-27 08:02:29 +0000 UTC" firstStartedPulling="2025-12-27 08:02:40.353563163 +0000 UTC m=+1221.174003757" lastFinishedPulling="2025-12-27 08:02:57.841296582 +0000 UTC m=+1238.661737176" observedRunningTime="2025-12-27 08:03:00.07617853 +0000 UTC m=+1240.896619144" watchObservedRunningTime="2025-12-27 08:03:00.079732789 +0000 UTC m=+1240.900173383" Dec 27 08:03:00 crc kubenswrapper[4934]: I1227 08:03:00.084152 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 27 08:03:00 crc kubenswrapper[4934]: I1227 08:03:00.113326 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-xp7tf" podStartSLOduration=9.113306291 podStartE2EDuration="9.113306291s" podCreationTimestamp="2025-12-27 08:02:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:03:00.101549216 +0000 UTC m=+1240.921989830" watchObservedRunningTime="2025-12-27 08:03:00.113306291 +0000 UTC m=+1240.933746885" Dec 27 08:03:00 crc kubenswrapper[4934]: I1227 08:03:00.129893 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-5v4n5" podStartSLOduration=4.623097108 podStartE2EDuration="10.129867566s" podCreationTimestamp="2025-12-27 08:02:50 +0000 UTC" firstStartedPulling="2025-12-27 08:02:52.361627423 +0000 UTC m=+1233.182068017" lastFinishedPulling="2025-12-27 08:02:57.868397881 +0000 UTC m=+1238.688838475" observedRunningTime="2025-12-27 08:03:00.117390503 +0000 UTC m=+1240.937831107" watchObservedRunningTime="2025-12-27 08:03:00.129867566 +0000 UTC m=+1240.950308160" Dec 27 08:03:00 crc kubenswrapper[4934]: I1227 08:03:00.159428 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=17.254685031 podStartE2EDuration="34.159410837s" podCreationTimestamp="2025-12-27 08:02:26 +0000 UTC" firstStartedPulling="2025-12-27 08:02:40.960292821 +0000 UTC m=+1221.780733415" lastFinishedPulling="2025-12-27 08:02:57.865018627 +0000 UTC m=+1238.685459221" observedRunningTime="2025-12-27 08:03:00.155625102 +0000 UTC m=+1240.976065726" watchObservedRunningTime="2025-12-27 08:03:00.159410837 +0000 UTC m=+1240.979851431" Dec 27 08:03:00 crc kubenswrapper[4934]: I1227 08:03:00.167435 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 27 08:03:01 crc kubenswrapper[4934]: I1227 08:03:01.109181 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 27 08:03:01 crc kubenswrapper[4934]: I1227 08:03:01.116299 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 27 08:03:01 crc kubenswrapper[4934]: I1227 08:03:01.116353 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 27 08:03:01 crc kubenswrapper[4934]: I1227 08:03:01.160936 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 27 08:03:01 crc kubenswrapper[4934]: I1227 08:03:01.161356 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 27 08:03:01 crc kubenswrapper[4934]: I1227 08:03:01.482658 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-g2q82"] Dec 27 08:03:01 crc kubenswrapper[4934]: E1227 08:03:01.483018 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbe083f3-5a60-4b98-b13f-65b82fcae1c1" containerName="init" Dec 27 08:03:01 crc kubenswrapper[4934]: I1227 08:03:01.483036 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbe083f3-5a60-4b98-b13f-65b82fcae1c1" containerName="init" Dec 27 08:03:01 crc kubenswrapper[4934]: I1227 08:03:01.483237 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbe083f3-5a60-4b98-b13f-65b82fcae1c1" containerName="init" Dec 27 08:03:01 crc kubenswrapper[4934]: I1227 08:03:01.483925 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g2q82" Dec 27 08:03:01 crc kubenswrapper[4934]: I1227 08:03:01.494547 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-g2q82"] Dec 27 08:03:01 crc kubenswrapper[4934]: I1227 08:03:01.568405 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czmln\" (UniqueName: \"kubernetes.io/projected/956ddca3-cc80-4e88-8695-4d18f77c9292-kube-api-access-czmln\") pod \"keystone-db-create-g2q82\" (UID: \"956ddca3-cc80-4e88-8695-4d18f77c9292\") " pod="openstack/keystone-db-create-g2q82" Dec 27 08:03:01 crc kubenswrapper[4934]: I1227 08:03:01.670638 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czmln\" (UniqueName: \"kubernetes.io/projected/956ddca3-cc80-4e88-8695-4d18f77c9292-kube-api-access-czmln\") pod \"keystone-db-create-g2q82\" (UID: \"956ddca3-cc80-4e88-8695-4d18f77c9292\") " pod="openstack/keystone-db-create-g2q82" Dec 27 08:03:01 crc kubenswrapper[4934]: I1227 08:03:01.712616 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-29dwl"] Dec 27 08:03:01 crc kubenswrapper[4934]: I1227 08:03:01.715770 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-29dwl" Dec 27 08:03:01 crc kubenswrapper[4934]: I1227 08:03:01.727631 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czmln\" (UniqueName: \"kubernetes.io/projected/956ddca3-cc80-4e88-8695-4d18f77c9292-kube-api-access-czmln\") pod \"keystone-db-create-g2q82\" (UID: \"956ddca3-cc80-4e88-8695-4d18f77c9292\") " pod="openstack/keystone-db-create-g2q82" Dec 27 08:03:01 crc kubenswrapper[4934]: I1227 08:03:01.728001 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-29dwl"] Dec 27 08:03:01 crc kubenswrapper[4934]: I1227 08:03:01.772453 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmvcv\" (UniqueName: \"kubernetes.io/projected/c0ade00b-01f5-493e-89ab-1925176e3bf0-kube-api-access-zmvcv\") pod \"placement-db-create-29dwl\" (UID: \"c0ade00b-01f5-493e-89ab-1925176e3bf0\") " pod="openstack/placement-db-create-29dwl" Dec 27 08:03:01 crc kubenswrapper[4934]: I1227 08:03:01.828902 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g2q82" Dec 27 08:03:01 crc kubenswrapper[4934]: I1227 08:03:01.873961 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmvcv\" (UniqueName: \"kubernetes.io/projected/c0ade00b-01f5-493e-89ab-1925176e3bf0-kube-api-access-zmvcv\") pod \"placement-db-create-29dwl\" (UID: \"c0ade00b-01f5-493e-89ab-1925176e3bf0\") " pod="openstack/placement-db-create-29dwl" Dec 27 08:03:01 crc kubenswrapper[4934]: I1227 08:03:01.896265 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmvcv\" (UniqueName: \"kubernetes.io/projected/c0ade00b-01f5-493e-89ab-1925176e3bf0-kube-api-access-zmvcv\") pod \"placement-db-create-29dwl\" (UID: \"c0ade00b-01f5-493e-89ab-1925176e3bf0\") " pod="openstack/placement-db-create-29dwl" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.087455 4934 generic.go:334] "Generic (PLEG): container finished" podID="92d66221-653f-4837-a733-f02073b6ed70" containerID="6984265a9dcba06e22bdc66f68ad44c0acc9f56b5319d26f0929c8d549b57ab3" exitCode=0 Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.088182 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92d66221-653f-4837-a733-f02073b6ed70","Type":"ContainerDied","Data":"6984265a9dcba06e22bdc66f68ad44c0acc9f56b5319d26f0929c8d549b57ab3"} Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.088796 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-29dwl" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.091556 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.092017 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.166912 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.177664 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.338014 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-g2q82"] Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.560951 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.566002 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.571861 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.590336 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.590587 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.590718 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-nv46z" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.592836 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/727f882e-fa63-46e9-9cd5-9781e472f2ef-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"727f882e-fa63-46e9-9cd5-9781e472f2ef\") " pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.592928 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/727f882e-fa63-46e9-9cd5-9781e472f2ef-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"727f882e-fa63-46e9-9cd5-9781e472f2ef\") " pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.593015 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/727f882e-fa63-46e9-9cd5-9781e472f2ef-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"727f882e-fa63-46e9-9cd5-9781e472f2ef\") " pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.593072 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/727f882e-fa63-46e9-9cd5-9781e472f2ef-config\") pod \"ovn-northd-0\" (UID: \"727f882e-fa63-46e9-9cd5-9781e472f2ef\") " pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.593506 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/727f882e-fa63-46e9-9cd5-9781e472f2ef-scripts\") pod \"ovn-northd-0\" (UID: \"727f882e-fa63-46e9-9cd5-9781e472f2ef\") " pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.593543 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mknb6\" (UniqueName: \"kubernetes.io/projected/727f882e-fa63-46e9-9cd5-9781e472f2ef-kube-api-access-mknb6\") pod \"ovn-northd-0\" (UID: \"727f882e-fa63-46e9-9cd5-9781e472f2ef\") " pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.593571 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/727f882e-fa63-46e9-9cd5-9781e472f2ef-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"727f882e-fa63-46e9-9cd5-9781e472f2ef\") " pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.594343 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.655729 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-29dwl"] Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.695022 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/727f882e-fa63-46e9-9cd5-9781e472f2ef-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"727f882e-fa63-46e9-9cd5-9781e472f2ef\") " pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.695114 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/727f882e-fa63-46e9-9cd5-9781e472f2ef-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"727f882e-fa63-46e9-9cd5-9781e472f2ef\") " pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.695180 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/727f882e-fa63-46e9-9cd5-9781e472f2ef-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"727f882e-fa63-46e9-9cd5-9781e472f2ef\") " pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.695257 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/727f882e-fa63-46e9-9cd5-9781e472f2ef-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"727f882e-fa63-46e9-9cd5-9781e472f2ef\") " pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.695285 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/727f882e-fa63-46e9-9cd5-9781e472f2ef-config\") pod \"ovn-northd-0\" (UID: \"727f882e-fa63-46e9-9cd5-9781e472f2ef\") " pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.695321 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/727f882e-fa63-46e9-9cd5-9781e472f2ef-scripts\") pod \"ovn-northd-0\" (UID: \"727f882e-fa63-46e9-9cd5-9781e472f2ef\") " pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.695342 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mknb6\" (UniqueName: \"kubernetes.io/projected/727f882e-fa63-46e9-9cd5-9781e472f2ef-kube-api-access-mknb6\") pod \"ovn-northd-0\" (UID: \"727f882e-fa63-46e9-9cd5-9781e472f2ef\") " pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.696611 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/727f882e-fa63-46e9-9cd5-9781e472f2ef-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"727f882e-fa63-46e9-9cd5-9781e472f2ef\") " pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.696972 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/727f882e-fa63-46e9-9cd5-9781e472f2ef-scripts\") pod \"ovn-northd-0\" (UID: \"727f882e-fa63-46e9-9cd5-9781e472f2ef\") " pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.698329 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/727f882e-fa63-46e9-9cd5-9781e472f2ef-config\") pod \"ovn-northd-0\" (UID: \"727f882e-fa63-46e9-9cd5-9781e472f2ef\") " pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.703499 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/727f882e-fa63-46e9-9cd5-9781e472f2ef-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"727f882e-fa63-46e9-9cd5-9781e472f2ef\") " pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.703064 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/727f882e-fa63-46e9-9cd5-9781e472f2ef-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"727f882e-fa63-46e9-9cd5-9781e472f2ef\") " pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.707694 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/727f882e-fa63-46e9-9cd5-9781e472f2ef-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"727f882e-fa63-46e9-9cd5-9781e472f2ef\") " pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.713393 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mknb6\" (UniqueName: \"kubernetes.io/projected/727f882e-fa63-46e9-9cd5-9781e472f2ef-kube-api-access-mknb6\") pod \"ovn-northd-0\" (UID: \"727f882e-fa63-46e9-9cd5-9781e472f2ef\") " pod="openstack/ovn-northd-0" Dec 27 08:03:02 crc kubenswrapper[4934]: I1227 08:03:02.913543 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 27 08:03:03 crc kubenswrapper[4934]: I1227 08:03:03.109204 4934 generic.go:334] "Generic (PLEG): container finished" podID="956ddca3-cc80-4e88-8695-4d18f77c9292" containerID="25069ee0e99208f73e7ad1aa0c872b35b7fddc1fe9590dda7a921889999be47d" exitCode=0 Dec 27 08:03:03 crc kubenswrapper[4934]: I1227 08:03:03.109715 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g2q82" event={"ID":"956ddca3-cc80-4e88-8695-4d18f77c9292","Type":"ContainerDied","Data":"25069ee0e99208f73e7ad1aa0c872b35b7fddc1fe9590dda7a921889999be47d"} Dec 27 08:03:03 crc kubenswrapper[4934]: I1227 08:03:03.109754 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g2q82" event={"ID":"956ddca3-cc80-4e88-8695-4d18f77c9292","Type":"ContainerStarted","Data":"833feaef8a1878dd4c7cd01fed1470feee9e4c80d13deaa65036a0a95b057421"} Dec 27 08:03:03 crc kubenswrapper[4934]: I1227 08:03:03.132780 4934 generic.go:334] "Generic (PLEG): container finished" podID="c0ade00b-01f5-493e-89ab-1925176e3bf0" containerID="d140ccfe6a91e2ba03a1ec5b8f9e5b1705101d5ce3274a3b70cf3771a3572cf2" exitCode=0 Dec 27 08:03:03 crc kubenswrapper[4934]: I1227 08:03:03.132954 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-29dwl" event={"ID":"c0ade00b-01f5-493e-89ab-1925176e3bf0","Type":"ContainerDied","Data":"d140ccfe6a91e2ba03a1ec5b8f9e5b1705101d5ce3274a3b70cf3771a3572cf2"} Dec 27 08:03:03 crc kubenswrapper[4934]: I1227 08:03:03.133037 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-29dwl" event={"ID":"c0ade00b-01f5-493e-89ab-1925176e3bf0","Type":"ContainerStarted","Data":"8e8ee7f73ac4cdf5fe4bff3a2d1a357583cb70a18c4cd2880ed7547f72335727"} Dec 27 08:03:03 crc kubenswrapper[4934]: I1227 08:03:03.800244 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-xp7tf"] Dec 27 08:03:03 crc kubenswrapper[4934]: I1227 08:03:03.800475 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-xp7tf" podUID="c9266772-ff58-4a27-ad46-99f8b6e344bd" containerName="dnsmasq-dns" containerID="cri-o://ca21fca8f6f9ffe20c122ba46c188e52c5a204d1c7b7d64b8d8308ee9eaad95b" gracePeriod=10 Dec 27 08:03:03 crc kubenswrapper[4934]: I1227 08:03:03.830044 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-5n5c8"] Dec 27 08:03:03 crc kubenswrapper[4934]: I1227 08:03:03.832035 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:03:03 crc kubenswrapper[4934]: I1227 08:03:03.852247 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 27 08:03:03 crc kubenswrapper[4934]: I1227 08:03:03.852835 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-5n5c8"] Dec 27 08:03:03 crc kubenswrapper[4934]: I1227 08:03:03.973302 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-wqrwd"] Dec 27 08:03:03 crc kubenswrapper[4934]: I1227 08:03:03.975569 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-wqrwd" Dec 27 08:03:03 crc kubenswrapper[4934]: I1227 08:03:03.999623 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-wqrwd"] Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.034348 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-5n5c8\" (UID: \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\") " pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.034420 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9lhz\" (UniqueName: \"kubernetes.io/projected/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-kube-api-access-z9lhz\") pod \"dnsmasq-dns-b8fbc5445-5n5c8\" (UID: \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\") " pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.034463 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-5n5c8\" (UID: \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\") " pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.034561 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-5n5c8\" (UID: \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\") " pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.034605 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-config\") pod \"dnsmasq-dns-b8fbc5445-5n5c8\" (UID: \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\") " pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.143870 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-5n5c8\" (UID: \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\") " pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.144225 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-5n5c8\" (UID: \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\") " pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.144353 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-config\") pod \"dnsmasq-dns-b8fbc5445-5n5c8\" (UID: \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\") " pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.144545 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2jpq\" (UniqueName: \"kubernetes.io/projected/a71b8c67-af1b-4c61-9820-fad0c957733a-kube-api-access-p2jpq\") pod \"mysqld-exporter-openstack-db-create-wqrwd\" (UID: \"a71b8c67-af1b-4c61-9820-fad0c957733a\") " pod="openstack/mysqld-exporter-openstack-db-create-wqrwd" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.144779 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-5n5c8\" (UID: \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\") " pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.144888 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9lhz\" (UniqueName: \"kubernetes.io/projected/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-kube-api-access-z9lhz\") pod \"dnsmasq-dns-b8fbc5445-5n5c8\" (UID: \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\") " pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.145345 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-5n5c8\" (UID: \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\") " pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.146238 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-5n5c8\" (UID: \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\") " pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.147250 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-5n5c8\" (UID: \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\") " pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.149878 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-config\") pod \"dnsmasq-dns-b8fbc5445-5n5c8\" (UID: \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\") " pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.171470 4934 generic.go:334] "Generic (PLEG): container finished" podID="c9266772-ff58-4a27-ad46-99f8b6e344bd" containerID="ca21fca8f6f9ffe20c122ba46c188e52c5a204d1c7b7d64b8d8308ee9eaad95b" exitCode=0 Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.172163 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9lhz\" (UniqueName: \"kubernetes.io/projected/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-kube-api-access-z9lhz\") pod \"dnsmasq-dns-b8fbc5445-5n5c8\" (UID: \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\") " pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.172280 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-xp7tf" event={"ID":"c9266772-ff58-4a27-ad46-99f8b6e344bd","Type":"ContainerDied","Data":"ca21fca8f6f9ffe20c122ba46c188e52c5a204d1c7b7d64b8d8308ee9eaad95b"} Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.204707 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.246385 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2jpq\" (UniqueName: \"kubernetes.io/projected/a71b8c67-af1b-4c61-9820-fad0c957733a-kube-api-access-p2jpq\") pod \"mysqld-exporter-openstack-db-create-wqrwd\" (UID: \"a71b8c67-af1b-4c61-9820-fad0c957733a\") " pod="openstack/mysqld-exporter-openstack-db-create-wqrwd" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.280181 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2jpq\" (UniqueName: \"kubernetes.io/projected/a71b8c67-af1b-4c61-9820-fad0c957733a-kube-api-access-p2jpq\") pod \"mysqld-exporter-openstack-db-create-wqrwd\" (UID: \"a71b8c67-af1b-4c61-9820-fad0c957733a\") " pod="openstack/mysqld-exporter-openstack-db-create-wqrwd" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.289636 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 27 08:03:04 crc kubenswrapper[4934]: W1227 08:03:04.295033 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod727f882e_fa63_46e9_9cd5_9781e472f2ef.slice/crio-11a2d4af452eeaca0f9d5c54da23bc908ea26b3b48ada9c8f61db8608aed2b52 WatchSource:0}: Error finding container 11a2d4af452eeaca0f9d5c54da23bc908ea26b3b48ada9c8f61db8608aed2b52: Status 404 returned error can't find the container with id 11a2d4af452eeaca0f9d5c54da23bc908ea26b3b48ada9c8f61db8608aed2b52 Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.296685 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-wqrwd" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.445248 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-xp7tf" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.562874 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmfh6\" (UniqueName: \"kubernetes.io/projected/c9266772-ff58-4a27-ad46-99f8b6e344bd-kube-api-access-xmfh6\") pod \"c9266772-ff58-4a27-ad46-99f8b6e344bd\" (UID: \"c9266772-ff58-4a27-ad46-99f8b6e344bd\") " Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.562954 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-ovsdbserver-sb\") pod \"c9266772-ff58-4a27-ad46-99f8b6e344bd\" (UID: \"c9266772-ff58-4a27-ad46-99f8b6e344bd\") " Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.563053 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-ovsdbserver-nb\") pod \"c9266772-ff58-4a27-ad46-99f8b6e344bd\" (UID: \"c9266772-ff58-4a27-ad46-99f8b6e344bd\") " Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.563113 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-config\") pod \"c9266772-ff58-4a27-ad46-99f8b6e344bd\" (UID: \"c9266772-ff58-4a27-ad46-99f8b6e344bd\") " Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.563154 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-dns-svc\") pod \"c9266772-ff58-4a27-ad46-99f8b6e344bd\" (UID: \"c9266772-ff58-4a27-ad46-99f8b6e344bd\") " Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.575759 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9266772-ff58-4a27-ad46-99f8b6e344bd-kube-api-access-xmfh6" (OuterVolumeSpecName: "kube-api-access-xmfh6") pod "c9266772-ff58-4a27-ad46-99f8b6e344bd" (UID: "c9266772-ff58-4a27-ad46-99f8b6e344bd"). InnerVolumeSpecName "kube-api-access-xmfh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.628630 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c9266772-ff58-4a27-ad46-99f8b6e344bd" (UID: "c9266772-ff58-4a27-ad46-99f8b6e344bd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.642146 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c9266772-ff58-4a27-ad46-99f8b6e344bd" (UID: "c9266772-ff58-4a27-ad46-99f8b6e344bd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.652261 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c9266772-ff58-4a27-ad46-99f8b6e344bd" (UID: "c9266772-ff58-4a27-ad46-99f8b6e344bd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.664459 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-config" (OuterVolumeSpecName: "config") pod "c9266772-ff58-4a27-ad46-99f8b6e344bd" (UID: "c9266772-ff58-4a27-ad46-99f8b6e344bd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.666582 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.666618 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.666630 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.666642 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmfh6\" (UniqueName: \"kubernetes.io/projected/c9266772-ff58-4a27-ad46-99f8b6e344bd-kube-api-access-xmfh6\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.666653 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c9266772-ff58-4a27-ad46-99f8b6e344bd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.693512 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-29dwl" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.819781 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g2q82" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.870347 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmvcv\" (UniqueName: \"kubernetes.io/projected/c0ade00b-01f5-493e-89ab-1925176e3bf0-kube-api-access-zmvcv\") pod \"c0ade00b-01f5-493e-89ab-1925176e3bf0\" (UID: \"c0ade00b-01f5-493e-89ab-1925176e3bf0\") " Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.873930 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0ade00b-01f5-493e-89ab-1925176e3bf0-kube-api-access-zmvcv" (OuterVolumeSpecName: "kube-api-access-zmvcv") pod "c0ade00b-01f5-493e-89ab-1925176e3bf0" (UID: "c0ade00b-01f5-493e-89ab-1925176e3bf0"). InnerVolumeSpecName "kube-api-access-zmvcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.916371 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-5n5c8"] Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.925386 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 27 08:03:04 crc kubenswrapper[4934]: E1227 08:03:04.925924 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9266772-ff58-4a27-ad46-99f8b6e344bd" containerName="init" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.925942 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9266772-ff58-4a27-ad46-99f8b6e344bd" containerName="init" Dec 27 08:03:04 crc kubenswrapper[4934]: E1227 08:03:04.925959 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9266772-ff58-4a27-ad46-99f8b6e344bd" containerName="dnsmasq-dns" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.925965 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9266772-ff58-4a27-ad46-99f8b6e344bd" containerName="dnsmasq-dns" Dec 27 08:03:04 crc kubenswrapper[4934]: E1227 08:03:04.925996 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="956ddca3-cc80-4e88-8695-4d18f77c9292" containerName="mariadb-database-create" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.926002 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="956ddca3-cc80-4e88-8695-4d18f77c9292" containerName="mariadb-database-create" Dec 27 08:03:04 crc kubenswrapper[4934]: E1227 08:03:04.926018 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0ade00b-01f5-493e-89ab-1925176e3bf0" containerName="mariadb-database-create" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.926024 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0ade00b-01f5-493e-89ab-1925176e3bf0" containerName="mariadb-database-create" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.926208 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="956ddca3-cc80-4e88-8695-4d18f77c9292" containerName="mariadb-database-create" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.926232 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9266772-ff58-4a27-ad46-99f8b6e344bd" containerName="dnsmasq-dns" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.926249 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0ade00b-01f5-493e-89ab-1925176e3bf0" containerName="mariadb-database-create" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.933187 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.937134 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.937183 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-xggzc" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.938627 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.938681 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.963139 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.974384 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czmln\" (UniqueName: \"kubernetes.io/projected/956ddca3-cc80-4e88-8695-4d18f77c9292-kube-api-access-czmln\") pod \"956ddca3-cc80-4e88-8695-4d18f77c9292\" (UID: \"956ddca3-cc80-4e88-8695-4d18f77c9292\") " Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.977715 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/956ddca3-cc80-4e88-8695-4d18f77c9292-kube-api-access-czmln" (OuterVolumeSpecName: "kube-api-access-czmln") pod "956ddca3-cc80-4e88-8695-4d18f77c9292" (UID: "956ddca3-cc80-4e88-8695-4d18f77c9292"). InnerVolumeSpecName "kube-api-access-czmln". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.978204 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmvcv\" (UniqueName: \"kubernetes.io/projected/c0ade00b-01f5-493e-89ab-1925176e3bf0-kube-api-access-zmvcv\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:04 crc kubenswrapper[4934]: I1227 08:03:04.978233 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czmln\" (UniqueName: \"kubernetes.io/projected/956ddca3-cc80-4e88-8695-4d18f77c9292-kube-api-access-czmln\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.079379 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b86q2\" (UniqueName: \"kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-kube-api-access-b86q2\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") " pod="openstack/swift-storage-0" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.079675 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e59489eb-aaa1-4188-b593-d3d3c2f1f412-lock\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") " pod="openstack/swift-storage-0" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.079711 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c4561afb-0c0b-42cc-81fd-db4cc7dfb27f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c4561afb-0c0b-42cc-81fd-db4cc7dfb27f\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") " pod="openstack/swift-storage-0" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.079731 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-etc-swift\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") " pod="openstack/swift-storage-0" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.080058 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e59489eb-aaa1-4188-b593-d3d3c2f1f412-cache\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") " pod="openstack/swift-storage-0" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.101954 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-wqrwd"] Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.182523 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e59489eb-aaa1-4188-b593-d3d3c2f1f412-lock\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") " pod="openstack/swift-storage-0" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.182581 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c4561afb-0c0b-42cc-81fd-db4cc7dfb27f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c4561afb-0c0b-42cc-81fd-db4cc7dfb27f\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") " pod="openstack/swift-storage-0" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.182602 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-etc-swift\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") " pod="openstack/swift-storage-0" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.182692 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e59489eb-aaa1-4188-b593-d3d3c2f1f412-cache\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") " pod="openstack/swift-storage-0" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.182722 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b86q2\" (UniqueName: \"kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-kube-api-access-b86q2\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") " pod="openstack/swift-storage-0" Dec 27 08:03:05 crc kubenswrapper[4934]: E1227 08:03:05.182963 4934 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 27 08:03:05 crc kubenswrapper[4934]: E1227 08:03:05.183001 4934 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 27 08:03:05 crc kubenswrapper[4934]: E1227 08:03:05.183097 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-etc-swift podName:e59489eb-aaa1-4188-b593-d3d3c2f1f412 nodeName:}" failed. No retries permitted until 2025-12-27 08:03:05.683060768 +0000 UTC m=+1246.503501372 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-etc-swift") pod "swift-storage-0" (UID: "e59489eb-aaa1-4188-b593-d3d3c2f1f412") : configmap "swift-ring-files" not found Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.183119 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e59489eb-aaa1-4188-b593-d3d3c2f1f412-lock\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") " pod="openstack/swift-storage-0" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.183334 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e59489eb-aaa1-4188-b593-d3d3c2f1f412-cache\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") " pod="openstack/swift-storage-0" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.191113 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.191151 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c4561afb-0c0b-42cc-81fd-db4cc7dfb27f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c4561afb-0c0b-42cc-81fd-db4cc7dfb27f\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e021005c537b584ba3f1055f8639eb1f4df7aaba0724c261d89e70c0a4aaaeb4/globalmount\"" pod="openstack/swift-storage-0" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.194803 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-g2q82" event={"ID":"956ddca3-cc80-4e88-8695-4d18f77c9292","Type":"ContainerDied","Data":"833feaef8a1878dd4c7cd01fed1470feee9e4c80d13deaa65036a0a95b057421"} Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.194825 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-g2q82" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.194847 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="833feaef8a1878dd4c7cd01fed1470feee9e4c80d13deaa65036a0a95b057421" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.197574 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-xp7tf" event={"ID":"c9266772-ff58-4a27-ad46-99f8b6e344bd","Type":"ContainerDied","Data":"2c0d3518c5e7906ff6664670cdf309af08307669e6446c4b4f9bf0318c9e9bc6"} Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.197621 4934 scope.go:117] "RemoveContainer" containerID="ca21fca8f6f9ffe20c122ba46c188e52c5a204d1c7b7d64b8d8308ee9eaad95b" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.197657 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-xp7tf" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.199679 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" event={"ID":"9ba80d2d-2587-4da1-aa21-31fc5e2775b0","Type":"ContainerStarted","Data":"af2309b0d8f07ac64a5d58a21e2725160bc2f158f8735417342fd085b404641e"} Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.201948 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-29dwl" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.201946 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-29dwl" event={"ID":"c0ade00b-01f5-493e-89ab-1925176e3bf0","Type":"ContainerDied","Data":"8e8ee7f73ac4cdf5fe4bff3a2d1a357583cb70a18c4cd2880ed7547f72335727"} Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.202049 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e8ee7f73ac4cdf5fe4bff3a2d1a357583cb70a18c4cd2880ed7547f72335727" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.204606 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"727f882e-fa63-46e9-9cd5-9781e472f2ef","Type":"ContainerStarted","Data":"11a2d4af452eeaca0f9d5c54da23bc908ea26b3b48ada9c8f61db8608aed2b52"} Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.209616 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b86q2\" (UniqueName: \"kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-kube-api-access-b86q2\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") " pod="openstack/swift-storage-0" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.212860 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-wqrwd" event={"ID":"a71b8c67-af1b-4c61-9820-fad0c957733a","Type":"ContainerStarted","Data":"bc1118d5c147ea58b6902d75acc1f6c611889ed1151b40bae9fe71e261df46ee"} Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.232832 4934 scope.go:117] "RemoveContainer" containerID="c4ef3840320e11fd8cd6f5eae183bb52ed263737f8d8dfb44351ea6b255802a8" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.258875 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c4561afb-0c0b-42cc-81fd-db4cc7dfb27f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c4561afb-0c0b-42cc-81fd-db4cc7dfb27f\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") " pod="openstack/swift-storage-0" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.268373 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-xp7tf"] Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.280192 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-xp7tf"] Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.447267 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-px8xc"] Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.450133 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.451723 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.464820 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.466496 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.488586 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9266772-ff58-4a27-ad46-99f8b6e344bd" path="/var/lib/kubelet/pods/c9266772-ff58-4a27-ad46-99f8b6e344bd/volumes" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.489204 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-px8xc"] Dec 27 08:03:05 crc kubenswrapper[4934]: E1227 08:03:05.489675 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-fhw5v ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-fhw5v ring-data-devices scripts swiftconf]: context canceled" pod="openstack/swift-ring-rebalance-px8xc" podUID="f387d624-e02c-4b40-b9c8-d8c575fc7a20" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.515652 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-bswkq"] Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.517172 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.586120 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bswkq"] Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.591510 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhw5v\" (UniqueName: \"kubernetes.io/projected/f387d624-e02c-4b40-b9c8-d8c575fc7a20-kube-api-access-fhw5v\") pod \"swift-ring-rebalance-px8xc\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.591559 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f387d624-e02c-4b40-b9c8-d8c575fc7a20-scripts\") pod \"swift-ring-rebalance-px8xc\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.591616 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f387d624-e02c-4b40-b9c8-d8c575fc7a20-dispersionconf\") pod \"swift-ring-rebalance-px8xc\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.591686 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f387d624-e02c-4b40-b9c8-d8c575fc7a20-swiftconf\") pod \"swift-ring-rebalance-px8xc\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.591715 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f387d624-e02c-4b40-b9c8-d8c575fc7a20-combined-ca-bundle\") pod \"swift-ring-rebalance-px8xc\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.591801 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f387d624-e02c-4b40-b9c8-d8c575fc7a20-etc-swift\") pod \"swift-ring-rebalance-px8xc\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.591828 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f387d624-e02c-4b40-b9c8-d8c575fc7a20-ring-data-devices\") pod \"swift-ring-rebalance-px8xc\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.607807 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-px8xc"] Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.694299 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f387d624-e02c-4b40-b9c8-d8c575fc7a20-dispersionconf\") pod \"swift-ring-rebalance-px8xc\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.694375 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-swiftconf\") pod \"swift-ring-rebalance-bswkq\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.694402 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-combined-ca-bundle\") pod \"swift-ring-rebalance-bswkq\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.694442 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f387d624-e02c-4b40-b9c8-d8c575fc7a20-swiftconf\") pod \"swift-ring-rebalance-px8xc\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.694469 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f387d624-e02c-4b40-b9c8-d8c575fc7a20-combined-ca-bundle\") pod \"swift-ring-rebalance-px8xc\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.694506 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-etc-swift\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") " pod="openstack/swift-storage-0" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.694529 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f387d624-e02c-4b40-b9c8-d8c575fc7a20-etc-swift\") pod \"swift-ring-rebalance-px8xc\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.694545 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f387d624-e02c-4b40-b9c8-d8c575fc7a20-ring-data-devices\") pod \"swift-ring-rebalance-px8xc\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.694588 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-dispersionconf\") pod \"swift-ring-rebalance-bswkq\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.694619 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-ring-data-devices\") pod \"swift-ring-rebalance-bswkq\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.694644 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-etc-swift\") pod \"swift-ring-rebalance-bswkq\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.694663 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tv65\" (UniqueName: \"kubernetes.io/projected/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-kube-api-access-6tv65\") pod \"swift-ring-rebalance-bswkq\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.694680 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhw5v\" (UniqueName: \"kubernetes.io/projected/f387d624-e02c-4b40-b9c8-d8c575fc7a20-kube-api-access-fhw5v\") pod \"swift-ring-rebalance-px8xc\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.694703 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f387d624-e02c-4b40-b9c8-d8c575fc7a20-scripts\") pod \"swift-ring-rebalance-px8xc\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.694747 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-scripts\") pod \"swift-ring-rebalance-bswkq\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: E1227 08:03:05.695704 4934 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 27 08:03:05 crc kubenswrapper[4934]: E1227 08:03:05.695738 4934 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 27 08:03:05 crc kubenswrapper[4934]: E1227 08:03:05.695797 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-etc-swift podName:e59489eb-aaa1-4188-b593-d3d3c2f1f412 nodeName:}" failed. No retries permitted until 2025-12-27 08:03:06.695774218 +0000 UTC m=+1247.516214872 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-etc-swift") pod "swift-storage-0" (UID: "e59489eb-aaa1-4188-b593-d3d3c2f1f412") : configmap "swift-ring-files" not found Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.696061 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f387d624-e02c-4b40-b9c8-d8c575fc7a20-ring-data-devices\") pod \"swift-ring-rebalance-px8xc\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.696300 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f387d624-e02c-4b40-b9c8-d8c575fc7a20-etc-swift\") pod \"swift-ring-rebalance-px8xc\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.697354 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f387d624-e02c-4b40-b9c8-d8c575fc7a20-scripts\") pod \"swift-ring-rebalance-px8xc\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.702920 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f387d624-e02c-4b40-b9c8-d8c575fc7a20-combined-ca-bundle\") pod \"swift-ring-rebalance-px8xc\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.703784 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f387d624-e02c-4b40-b9c8-d8c575fc7a20-swiftconf\") pod \"swift-ring-rebalance-px8xc\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.715054 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f387d624-e02c-4b40-b9c8-d8c575fc7a20-dispersionconf\") pod \"swift-ring-rebalance-px8xc\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.716030 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhw5v\" (UniqueName: \"kubernetes.io/projected/f387d624-e02c-4b40-b9c8-d8c575fc7a20-kube-api-access-fhw5v\") pod \"swift-ring-rebalance-px8xc\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.799423 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-swiftconf\") pod \"swift-ring-rebalance-bswkq\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.799471 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-combined-ca-bundle\") pod \"swift-ring-rebalance-bswkq\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.799636 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-dispersionconf\") pod \"swift-ring-rebalance-bswkq\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.799676 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-ring-data-devices\") pod \"swift-ring-rebalance-bswkq\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.799705 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-etc-swift\") pod \"swift-ring-rebalance-bswkq\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.799727 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tv65\" (UniqueName: \"kubernetes.io/projected/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-kube-api-access-6tv65\") pod \"swift-ring-rebalance-bswkq\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.799796 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-scripts\") pod \"swift-ring-rebalance-bswkq\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.800612 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-scripts\") pod \"swift-ring-rebalance-bswkq\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.800705 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-etc-swift\") pod \"swift-ring-rebalance-bswkq\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.801163 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-ring-data-devices\") pod \"swift-ring-rebalance-bswkq\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.804507 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-combined-ca-bundle\") pod \"swift-ring-rebalance-bswkq\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.804646 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-dispersionconf\") pod \"swift-ring-rebalance-bswkq\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.804895 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-swiftconf\") pod \"swift-ring-rebalance-bswkq\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.815440 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tv65\" (UniqueName: \"kubernetes.io/projected/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-kube-api-access-6tv65\") pod \"swift-ring-rebalance-bswkq\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:05 crc kubenswrapper[4934]: I1227 08:03:05.861551 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.223386 4934 generic.go:334] "Generic (PLEG): container finished" podID="a71b8c67-af1b-4c61-9820-fad0c957733a" containerID="ddb9e9c22e3038ed6d0a692612f9934cb1e02c84c16773de0c65c293948fbc13" exitCode=0 Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.223462 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-wqrwd" event={"ID":"a71b8c67-af1b-4c61-9820-fad0c957733a","Type":"ContainerDied","Data":"ddb9e9c22e3038ed6d0a692612f9934cb1e02c84c16773de0c65c293948fbc13"} Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.226911 4934 generic.go:334] "Generic (PLEG): container finished" podID="9ba80d2d-2587-4da1-aa21-31fc5e2775b0" containerID="dc5c4a8cb477b83fb27aae7ca0babf37748a40f4d0ab4db1c056699a80f9bd1c" exitCode=0 Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.226993 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.227713 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" event={"ID":"9ba80d2d-2587-4da1-aa21-31fc5e2775b0","Type":"ContainerDied","Data":"dc5c4a8cb477b83fb27aae7ca0babf37748a40f4d0ab4db1c056699a80f9bd1c"} Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.265297 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.363887 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-bswkq"] Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.413166 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f387d624-e02c-4b40-b9c8-d8c575fc7a20-scripts\") pod \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.413240 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f387d624-e02c-4b40-b9c8-d8c575fc7a20-swiftconf\") pod \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.413315 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhw5v\" (UniqueName: \"kubernetes.io/projected/f387d624-e02c-4b40-b9c8-d8c575fc7a20-kube-api-access-fhw5v\") pod \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.413340 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f387d624-e02c-4b40-b9c8-d8c575fc7a20-etc-swift\") pod \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.413552 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f387d624-e02c-4b40-b9c8-d8c575fc7a20-dispersionconf\") pod \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.413575 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f387d624-e02c-4b40-b9c8-d8c575fc7a20-ring-data-devices\") pod \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.413608 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f387d624-e02c-4b40-b9c8-d8c575fc7a20-combined-ca-bundle\") pod \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\" (UID: \"f387d624-e02c-4b40-b9c8-d8c575fc7a20\") " Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.413814 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f387d624-e02c-4b40-b9c8-d8c575fc7a20-scripts" (OuterVolumeSpecName: "scripts") pod "f387d624-e02c-4b40-b9c8-d8c575fc7a20" (UID: "f387d624-e02c-4b40-b9c8-d8c575fc7a20"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.414069 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f387d624-e02c-4b40-b9c8-d8c575fc7a20-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f387d624-e02c-4b40-b9c8-d8c575fc7a20" (UID: "f387d624-e02c-4b40-b9c8-d8c575fc7a20"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.415358 4934 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/f387d624-e02c-4b40-b9c8-d8c575fc7a20-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.415375 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f387d624-e02c-4b40-b9c8-d8c575fc7a20-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.416639 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f387d624-e02c-4b40-b9c8-d8c575fc7a20-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "f387d624-e02c-4b40-b9c8-d8c575fc7a20" (UID: "f387d624-e02c-4b40-b9c8-d8c575fc7a20"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.419354 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f387d624-e02c-4b40-b9c8-d8c575fc7a20-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "f387d624-e02c-4b40-b9c8-d8c575fc7a20" (UID: "f387d624-e02c-4b40-b9c8-d8c575fc7a20"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.421151 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f387d624-e02c-4b40-b9c8-d8c575fc7a20-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f387d624-e02c-4b40-b9c8-d8c575fc7a20" (UID: "f387d624-e02c-4b40-b9c8-d8c575fc7a20"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.423575 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f387d624-e02c-4b40-b9c8-d8c575fc7a20-kube-api-access-fhw5v" (OuterVolumeSpecName: "kube-api-access-fhw5v") pod "f387d624-e02c-4b40-b9c8-d8c575fc7a20" (UID: "f387d624-e02c-4b40-b9c8-d8c575fc7a20"). InnerVolumeSpecName "kube-api-access-fhw5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.424977 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f387d624-e02c-4b40-b9c8-d8c575fc7a20-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "f387d624-e02c-4b40-b9c8-d8c575fc7a20" (UID: "f387d624-e02c-4b40-b9c8-d8c575fc7a20"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.517004 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhw5v\" (UniqueName: \"kubernetes.io/projected/f387d624-e02c-4b40-b9c8-d8c575fc7a20-kube-api-access-fhw5v\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.517352 4934 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/f387d624-e02c-4b40-b9c8-d8c575fc7a20-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.517364 4934 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/f387d624-e02c-4b40-b9c8-d8c575fc7a20-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.517372 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f387d624-e02c-4b40-b9c8-d8c575fc7a20-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.517381 4934 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/f387d624-e02c-4b40-b9c8-d8c575fc7a20-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.721610 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-etc-swift\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") " pod="openstack/swift-storage-0" Dec 27 08:03:06 crc kubenswrapper[4934]: E1227 08:03:06.721820 4934 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 27 08:03:06 crc kubenswrapper[4934]: E1227 08:03:06.721856 4934 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 27 08:03:06 crc kubenswrapper[4934]: E1227 08:03:06.721924 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-etc-swift podName:e59489eb-aaa1-4188-b593-d3d3c2f1f412 nodeName:}" failed. No retries permitted until 2025-12-27 08:03:08.721903736 +0000 UTC m=+1249.542344330 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-etc-swift") pod "swift-storage-0" (UID: "e59489eb-aaa1-4188-b593-d3d3c2f1f412") : configmap "swift-ring-files" not found Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.959557 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-cgffw"] Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.961621 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cgffw" Dec 27 08:03:06 crc kubenswrapper[4934]: I1227 08:03:06.984627 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-cgffw"] Dec 27 08:03:07 crc kubenswrapper[4934]: I1227 08:03:07.131671 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knpzf\" (UniqueName: \"kubernetes.io/projected/7ede7d3e-b268-4132-b5d3-70569e23f060-kube-api-access-knpzf\") pod \"glance-db-create-cgffw\" (UID: \"7ede7d3e-b268-4132-b5d3-70569e23f060\") " pod="openstack/glance-db-create-cgffw" Dec 27 08:03:07 crc kubenswrapper[4934]: I1227 08:03:07.235702 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knpzf\" (UniqueName: \"kubernetes.io/projected/7ede7d3e-b268-4132-b5d3-70569e23f060-kube-api-access-knpzf\") pod \"glance-db-create-cgffw\" (UID: \"7ede7d3e-b268-4132-b5d3-70569e23f060\") " pod="openstack/glance-db-create-cgffw" Dec 27 08:03:07 crc kubenswrapper[4934]: I1227 08:03:07.246512 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" event={"ID":"9ba80d2d-2587-4da1-aa21-31fc5e2775b0","Type":"ContainerStarted","Data":"6f020a4ec55892c99858e76f37cbc694ceac6da4d533f1f1836c0b62bee489f2"} Dec 27 08:03:07 crc kubenswrapper[4934]: I1227 08:03:07.247914 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:03:07 crc kubenswrapper[4934]: I1227 08:03:07.266146 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"727f882e-fa63-46e9-9cd5-9781e472f2ef","Type":"ContainerStarted","Data":"903945fce2b24bcb371c05480fc9cbb565d137a45e7cd96534a057196f20bb81"} Dec 27 08:03:07 crc kubenswrapper[4934]: I1227 08:03:07.266190 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"727f882e-fa63-46e9-9cd5-9781e472f2ef","Type":"ContainerStarted","Data":"4fd62a9d2411fbe1cab09b321dd5e1651718ca62d130f296f4dc801144e5abc7"} Dec 27 08:03:07 crc kubenswrapper[4934]: I1227 08:03:07.267169 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 27 08:03:07 crc kubenswrapper[4934]: I1227 08:03:07.267453 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" podStartSLOduration=4.26743627 podStartE2EDuration="4.26743627s" podCreationTimestamp="2025-12-27 08:03:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:03:07.267328967 +0000 UTC m=+1248.087769561" watchObservedRunningTime="2025-12-27 08:03:07.26743627 +0000 UTC m=+1248.087876864" Dec 27 08:03:07 crc kubenswrapper[4934]: I1227 08:03:07.279864 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knpzf\" (UniqueName: \"kubernetes.io/projected/7ede7d3e-b268-4132-b5d3-70569e23f060-kube-api-access-knpzf\") pod \"glance-db-create-cgffw\" (UID: \"7ede7d3e-b268-4132-b5d3-70569e23f060\") " pod="openstack/glance-db-create-cgffw" Dec 27 08:03:07 crc kubenswrapper[4934]: I1227 08:03:07.280607 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bswkq" event={"ID":"ccb12f0b-32d5-4ab4-9228-a8aa005b1304","Type":"ContainerStarted","Data":"7807c4891b8905cf95b8e80bc4e9874aeb5d16aea8fb3de2b4defa363d31f09f"} Dec 27 08:03:07 crc kubenswrapper[4934]: I1227 08:03:07.280782 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-px8xc" Dec 27 08:03:07 crc kubenswrapper[4934]: I1227 08:03:07.300951 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.239846872 podStartE2EDuration="5.30093334s" podCreationTimestamp="2025-12-27 08:03:02 +0000 UTC" firstStartedPulling="2025-12-27 08:03:04.313349682 +0000 UTC m=+1245.133790276" lastFinishedPulling="2025-12-27 08:03:06.37443615 +0000 UTC m=+1247.194876744" observedRunningTime="2025-12-27 08:03:07.286049607 +0000 UTC m=+1248.106490201" watchObservedRunningTime="2025-12-27 08:03:07.30093334 +0000 UTC m=+1248.121373934" Dec 27 08:03:07 crc kubenswrapper[4934]: I1227 08:03:07.301070 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cgffw" Dec 27 08:03:07 crc kubenswrapper[4934]: I1227 08:03:07.357201 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-px8xc"] Dec 27 08:03:07 crc kubenswrapper[4934]: I1227 08:03:07.360300 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-px8xc"] Dec 27 08:03:07 crc kubenswrapper[4934]: I1227 08:03:07.481570 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f387d624-e02c-4b40-b9c8-d8c575fc7a20" path="/var/lib/kubelet/pods/f387d624-e02c-4b40-b9c8-d8c575fc7a20/volumes" Dec 27 08:03:07 crc kubenswrapper[4934]: I1227 08:03:07.710110 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-wqrwd" Dec 27 08:03:07 crc kubenswrapper[4934]: I1227 08:03:07.856312 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2jpq\" (UniqueName: \"kubernetes.io/projected/a71b8c67-af1b-4c61-9820-fad0c957733a-kube-api-access-p2jpq\") pod \"a71b8c67-af1b-4c61-9820-fad0c957733a\" (UID: \"a71b8c67-af1b-4c61-9820-fad0c957733a\") " Dec 27 08:03:07 crc kubenswrapper[4934]: I1227 08:03:07.875194 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a71b8c67-af1b-4c61-9820-fad0c957733a-kube-api-access-p2jpq" (OuterVolumeSpecName: "kube-api-access-p2jpq") pod "a71b8c67-af1b-4c61-9820-fad0c957733a" (UID: "a71b8c67-af1b-4c61-9820-fad0c957733a"). InnerVolumeSpecName "kube-api-access-p2jpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:07 crc kubenswrapper[4934]: I1227 08:03:07.959778 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2jpq\" (UniqueName: \"kubernetes.io/projected/a71b8c67-af1b-4c61-9820-fad0c957733a-kube-api-access-p2jpq\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:08 crc kubenswrapper[4934]: I1227 08:03:08.060217 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-cgffw"] Dec 27 08:03:08 crc kubenswrapper[4934]: I1227 08:03:08.293447 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-wqrwd" event={"ID":"a71b8c67-af1b-4c61-9820-fad0c957733a","Type":"ContainerDied","Data":"bc1118d5c147ea58b6902d75acc1f6c611889ed1151b40bae9fe71e261df46ee"} Dec 27 08:03:08 crc kubenswrapper[4934]: I1227 08:03:08.293494 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc1118d5c147ea58b6902d75acc1f6c611889ed1151b40bae9fe71e261df46ee" Dec 27 08:03:08 crc kubenswrapper[4934]: I1227 08:03:08.293652 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-wqrwd" Dec 27 08:03:08 crc kubenswrapper[4934]: I1227 08:03:08.779242 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-etc-swift\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") " pod="openstack/swift-storage-0" Dec 27 08:03:08 crc kubenswrapper[4934]: E1227 08:03:08.779710 4934 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 27 08:03:08 crc kubenswrapper[4934]: E1227 08:03:08.779731 4934 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 27 08:03:08 crc kubenswrapper[4934]: E1227 08:03:08.779778 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-etc-swift podName:e59489eb-aaa1-4188-b593-d3d3c2f1f412 nodeName:}" failed. No retries permitted until 2025-12-27 08:03:12.779762774 +0000 UTC m=+1253.600203368 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-etc-swift") pod "swift-storage-0" (UID: "e59489eb-aaa1-4188-b593-d3d3c2f1f412") : configmap "swift-ring-files" not found Dec 27 08:03:10 crc kubenswrapper[4934]: W1227 08:03:10.266239 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ede7d3e_b268_4132_b5d3_70569e23f060.slice/crio-72350978495f998b78cb3e951e8951247640c946c6e07cfd3933954fd342fb9e WatchSource:0}: Error finding container 72350978495f998b78cb3e951e8951247640c946c6e07cfd3933954fd342fb9e: Status 404 returned error can't find the container with id 72350978495f998b78cb3e951e8951247640c946c6e07cfd3933954fd342fb9e Dec 27 08:03:10 crc kubenswrapper[4934]: I1227 08:03:10.334221 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-cgffw" event={"ID":"7ede7d3e-b268-4132-b5d3-70569e23f060","Type":"ContainerStarted","Data":"72350978495f998b78cb3e951e8951247640c946c6e07cfd3933954fd342fb9e"} Dec 27 08:03:10 crc kubenswrapper[4934]: I1227 08:03:10.900265 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-86b5d5c74c-c9lp9" podUID="8aa711d1-ae51-480c-aa8b-9a510dfd7b6a" containerName="console" containerID="cri-o://eacd15e000e9c4e9be81a96aad383398017a9643d62d695c37cfa5fb8b300d93" gracePeriod=15 Dec 27 08:03:11 crc kubenswrapper[4934]: I1227 08:03:11.345292 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-86b5d5c74c-c9lp9_8aa711d1-ae51-480c-aa8b-9a510dfd7b6a/console/0.log" Dec 27 08:03:11 crc kubenswrapper[4934]: I1227 08:03:11.345584 4934 generic.go:334] "Generic (PLEG): container finished" podID="8aa711d1-ae51-480c-aa8b-9a510dfd7b6a" containerID="eacd15e000e9c4e9be81a96aad383398017a9643d62d695c37cfa5fb8b300d93" exitCode=2 Dec 27 08:03:11 crc kubenswrapper[4934]: I1227 08:03:11.345619 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-86b5d5c74c-c9lp9" event={"ID":"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a","Type":"ContainerDied","Data":"eacd15e000e9c4e9be81a96aad383398017a9643d62d695c37cfa5fb8b300d93"} Dec 27 08:03:11 crc kubenswrapper[4934]: I1227 08:03:11.520166 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-f9d8-account-create-dxnbx"] Dec 27 08:03:11 crc kubenswrapper[4934]: E1227 08:03:11.520663 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a71b8c67-af1b-4c61-9820-fad0c957733a" containerName="mariadb-database-create" Dec 27 08:03:11 crc kubenswrapper[4934]: I1227 08:03:11.520680 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a71b8c67-af1b-4c61-9820-fad0c957733a" containerName="mariadb-database-create" Dec 27 08:03:11 crc kubenswrapper[4934]: I1227 08:03:11.520972 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a71b8c67-af1b-4c61-9820-fad0c957733a" containerName="mariadb-database-create" Dec 27 08:03:11 crc kubenswrapper[4934]: I1227 08:03:11.521837 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f9d8-account-create-dxnbx" Dec 27 08:03:11 crc kubenswrapper[4934]: I1227 08:03:11.524842 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 27 08:03:11 crc kubenswrapper[4934]: I1227 08:03:11.534037 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f9d8-account-create-dxnbx"] Dec 27 08:03:11 crc kubenswrapper[4934]: I1227 08:03:11.657150 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l99th\" (UniqueName: \"kubernetes.io/projected/84b8ef67-1fd1-4d7c-82d3-cde56ca1468a-kube-api-access-l99th\") pod \"keystone-f9d8-account-create-dxnbx\" (UID: \"84b8ef67-1fd1-4d7c-82d3-cde56ca1468a\") " pod="openstack/keystone-f9d8-account-create-dxnbx" Dec 27 08:03:11 crc kubenswrapper[4934]: I1227 08:03:11.711434 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-f752-account-create-mxw52"] Dec 27 08:03:11 crc kubenswrapper[4934]: I1227 08:03:11.712853 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f752-account-create-mxw52" Dec 27 08:03:11 crc kubenswrapper[4934]: I1227 08:03:11.715877 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 27 08:03:11 crc kubenswrapper[4934]: I1227 08:03:11.734727 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f752-account-create-mxw52"] Dec 27 08:03:11 crc kubenswrapper[4934]: I1227 08:03:11.758433 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l99th\" (UniqueName: \"kubernetes.io/projected/84b8ef67-1fd1-4d7c-82d3-cde56ca1468a-kube-api-access-l99th\") pod \"keystone-f9d8-account-create-dxnbx\" (UID: \"84b8ef67-1fd1-4d7c-82d3-cde56ca1468a\") " pod="openstack/keystone-f9d8-account-create-dxnbx" Dec 27 08:03:11 crc kubenswrapper[4934]: I1227 08:03:11.775040 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l99th\" (UniqueName: \"kubernetes.io/projected/84b8ef67-1fd1-4d7c-82d3-cde56ca1468a-kube-api-access-l99th\") pod \"keystone-f9d8-account-create-dxnbx\" (UID: \"84b8ef67-1fd1-4d7c-82d3-cde56ca1468a\") " pod="openstack/keystone-f9d8-account-create-dxnbx" Dec 27 08:03:11 crc kubenswrapper[4934]: I1227 08:03:11.844382 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f9d8-account-create-dxnbx" Dec 27 08:03:11 crc kubenswrapper[4934]: I1227 08:03:11.860073 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnrnd\" (UniqueName: \"kubernetes.io/projected/766e151c-8962-4bd5-b174-4e381730ff6f-kube-api-access-dnrnd\") pod \"placement-f752-account-create-mxw52\" (UID: \"766e151c-8962-4bd5-b174-4e381730ff6f\") " pod="openstack/placement-f752-account-create-mxw52" Dec 27 08:03:11 crc kubenswrapper[4934]: I1227 08:03:11.963421 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnrnd\" (UniqueName: \"kubernetes.io/projected/766e151c-8962-4bd5-b174-4e381730ff6f-kube-api-access-dnrnd\") pod \"placement-f752-account-create-mxw52\" (UID: \"766e151c-8962-4bd5-b174-4e381730ff6f\") " pod="openstack/placement-f752-account-create-mxw52" Dec 27 08:03:12 crc kubenswrapper[4934]: I1227 08:03:12.002128 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnrnd\" (UniqueName: \"kubernetes.io/projected/766e151c-8962-4bd5-b174-4e381730ff6f-kube-api-access-dnrnd\") pod \"placement-f752-account-create-mxw52\" (UID: \"766e151c-8962-4bd5-b174-4e381730ff6f\") " pod="openstack/placement-f752-account-create-mxw52" Dec 27 08:03:12 crc kubenswrapper[4934]: I1227 08:03:12.036067 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f752-account-create-mxw52" Dec 27 08:03:12 crc kubenswrapper[4934]: I1227 08:03:12.259957 4934 patch_prober.go:28] interesting pod/console-86b5d5c74c-c9lp9 container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.88:8443/health\": dial tcp 10.217.0.88:8443: connect: connection refused" start-of-body= Dec 27 08:03:12 crc kubenswrapper[4934]: I1227 08:03:12.260026 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-86b5d5c74c-c9lp9" podUID="8aa711d1-ae51-480c-aa8b-9a510dfd7b6a" containerName="console" probeResult="failure" output="Get \"https://10.217.0.88:8443/health\": dial tcp 10.217.0.88:8443: connect: connection refused" Dec 27 08:03:12 crc kubenswrapper[4934]: I1227 08:03:12.357009 4934 generic.go:334] "Generic (PLEG): container finished" podID="5788f6c6-2500-4c66-af7a-830cde17caa4" containerID="025c6e94fc16f8dfe3303fd225b5fc2e02f9952414fc805476597fdb089f791b" exitCode=0 Dec 27 08:03:12 crc kubenswrapper[4934]: I1227 08:03:12.357313 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"5788f6c6-2500-4c66-af7a-830cde17caa4","Type":"ContainerDied","Data":"025c6e94fc16f8dfe3303fd225b5fc2e02f9952414fc805476597fdb089f791b"} Dec 27 08:03:12 crc kubenswrapper[4934]: I1227 08:03:12.361902 4934 generic.go:334] "Generic (PLEG): container finished" podID="df7d6b06-cb87-4635-8aca-1ab00a564dd7" containerID="6386108486d5f92741eb22a890f1a957ebb07da8c2f70cbbed654caafe93c1c2" exitCode=0 Dec 27 08:03:12 crc kubenswrapper[4934]: I1227 08:03:12.361955 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"df7d6b06-cb87-4635-8aca-1ab00a564dd7","Type":"ContainerDied","Data":"6386108486d5f92741eb22a890f1a957ebb07da8c2f70cbbed654caafe93c1c2"} Dec 27 08:03:12 crc kubenswrapper[4934]: I1227 08:03:12.782509 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-etc-swift\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") " pod="openstack/swift-storage-0" Dec 27 08:03:12 crc kubenswrapper[4934]: E1227 08:03:12.782699 4934 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 27 08:03:12 crc kubenswrapper[4934]: E1227 08:03:12.782975 4934 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 27 08:03:12 crc kubenswrapper[4934]: E1227 08:03:12.783034 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-etc-swift podName:e59489eb-aaa1-4188-b593-d3d3c2f1f412 nodeName:}" failed. No retries permitted until 2025-12-27 08:03:20.783016949 +0000 UTC m=+1261.603457543 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-etc-swift") pod "swift-storage-0" (UID: "e59489eb-aaa1-4188-b593-d3d3c2f1f412") : configmap "swift-ring-files" not found Dec 27 08:03:13 crc kubenswrapper[4934]: I1227 08:03:13.868657 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-622b-account-create-rz2xx"] Dec 27 08:03:13 crc kubenswrapper[4934]: I1227 08:03:13.871152 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-622b-account-create-rz2xx" Dec 27 08:03:13 crc kubenswrapper[4934]: I1227 08:03:13.873315 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-db-secret" Dec 27 08:03:13 crc kubenswrapper[4934]: I1227 08:03:13.924647 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-622b-account-create-rz2xx"] Dec 27 08:03:13 crc kubenswrapper[4934]: I1227 08:03:13.946054 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvlhp\" (UniqueName: \"kubernetes.io/projected/4e6e9101-e025-4566-833b-fd29fa46ab79-kube-api-access-nvlhp\") pod \"mysqld-exporter-622b-account-create-rz2xx\" (UID: \"4e6e9101-e025-4566-833b-fd29fa46ab79\") " pod="openstack/mysqld-exporter-622b-account-create-rz2xx" Dec 27 08:03:14 crc kubenswrapper[4934]: I1227 08:03:14.048915 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvlhp\" (UniqueName: \"kubernetes.io/projected/4e6e9101-e025-4566-833b-fd29fa46ab79-kube-api-access-nvlhp\") pod \"mysqld-exporter-622b-account-create-rz2xx\" (UID: \"4e6e9101-e025-4566-833b-fd29fa46ab79\") " pod="openstack/mysqld-exporter-622b-account-create-rz2xx" Dec 27 08:03:14 crc kubenswrapper[4934]: I1227 08:03:14.070513 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvlhp\" (UniqueName: \"kubernetes.io/projected/4e6e9101-e025-4566-833b-fd29fa46ab79-kube-api-access-nvlhp\") pod \"mysqld-exporter-622b-account-create-rz2xx\" (UID: \"4e6e9101-e025-4566-833b-fd29fa46ab79\") " pod="openstack/mysqld-exporter-622b-account-create-rz2xx" Dec 27 08:03:14 crc kubenswrapper[4934]: I1227 08:03:14.207417 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:03:14 crc kubenswrapper[4934]: I1227 08:03:14.245921 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-622b-account-create-rz2xx" Dec 27 08:03:14 crc kubenswrapper[4934]: I1227 08:03:14.290059 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5fzqq"] Dec 27 08:03:14 crc kubenswrapper[4934]: I1227 08:03:14.290278 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" podUID="ca9c1959-3f6a-4c4d-ab56-b464625d92c8" containerName="dnsmasq-dns" containerID="cri-o://8ef7f2c1ecadd3836be5e7f839be16b66848904a88d51eb92ee6d74ee50d8f77" gracePeriod=10 Dec 27 08:03:15 crc kubenswrapper[4934]: I1227 08:03:15.394008 4934 generic.go:334] "Generic (PLEG): container finished" podID="76158d63-2099-4418-82a2-09af0432c1a3" containerID="e6252db8825d6b9fe04ede9d278e6c3c2fa0bc3d3cf385da4ee6ce5043dc6b19" exitCode=0 Dec 27 08:03:15 crc kubenswrapper[4934]: I1227 08:03:15.396204 4934 generic.go:334] "Generic (PLEG): container finished" podID="cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c" containerID="9e896bff961270fe6bc84ed6b42b91ba71952e4040cd6b4594ed7c4e6be8c873" exitCode=0 Dec 27 08:03:15 crc kubenswrapper[4934]: I1227 08:03:15.398618 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"76158d63-2099-4418-82a2-09af0432c1a3","Type":"ContainerDied","Data":"e6252db8825d6b9fe04ede9d278e6c3c2fa0bc3d3cf385da4ee6ce5043dc6b19"} Dec 27 08:03:15 crc kubenswrapper[4934]: I1227 08:03:15.398757 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c","Type":"ContainerDied","Data":"9e896bff961270fe6bc84ed6b42b91ba71952e4040cd6b4594ed7c4e6be8c873"} Dec 27 08:03:16 crc kubenswrapper[4934]: I1227 08:03:16.410938 4934 generic.go:334] "Generic (PLEG): container finished" podID="ca9c1959-3f6a-4c4d-ab56-b464625d92c8" containerID="8ef7f2c1ecadd3836be5e7f839be16b66848904a88d51eb92ee6d74ee50d8f77" exitCode=0 Dec 27 08:03:16 crc kubenswrapper[4934]: I1227 08:03:16.410979 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" event={"ID":"ca9c1959-3f6a-4c4d-ab56-b464625d92c8","Type":"ContainerDied","Data":"8ef7f2c1ecadd3836be5e7f839be16b66848904a88d51eb92ee6d74ee50d8f77"} Dec 27 08:03:16 crc kubenswrapper[4934]: I1227 08:03:16.818848 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-86b5d5c74c-c9lp9_8aa711d1-ae51-480c-aa8b-9a510dfd7b6a/console/0.log" Dec 27 08:03:16 crc kubenswrapper[4934]: I1227 08:03:16.819343 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 08:03:16 crc kubenswrapper[4934]: I1227 08:03:16.914728 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cj5rv\" (UniqueName: \"kubernetes.io/projected/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-kube-api-access-cj5rv\") pod \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " Dec 27 08:03:16 crc kubenswrapper[4934]: I1227 08:03:16.914973 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-console-serving-cert\") pod \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " Dec 27 08:03:16 crc kubenswrapper[4934]: I1227 08:03:16.915120 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-service-ca\") pod \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " Dec 27 08:03:16 crc kubenswrapper[4934]: I1227 08:03:16.915150 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-trusted-ca-bundle\") pod \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " Dec 27 08:03:16 crc kubenswrapper[4934]: I1227 08:03:16.915277 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-console-oauth-config\") pod \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " Dec 27 08:03:16 crc kubenswrapper[4934]: I1227 08:03:16.915358 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-oauth-serving-cert\") pod \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " Dec 27 08:03:16 crc kubenswrapper[4934]: I1227 08:03:16.915399 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-console-config\") pod \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\" (UID: \"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a\") " Dec 27 08:03:16 crc kubenswrapper[4934]: I1227 08:03:16.915901 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-service-ca" (OuterVolumeSpecName: "service-ca") pod "8aa711d1-ae51-480c-aa8b-9a510dfd7b6a" (UID: "8aa711d1-ae51-480c-aa8b-9a510dfd7b6a"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:03:16 crc kubenswrapper[4934]: I1227 08:03:16.916253 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-console-config" (OuterVolumeSpecName: "console-config") pod "8aa711d1-ae51-480c-aa8b-9a510dfd7b6a" (UID: "8aa711d1-ae51-480c-aa8b-9a510dfd7b6a"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:03:16 crc kubenswrapper[4934]: I1227 08:03:16.916589 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "8aa711d1-ae51-480c-aa8b-9a510dfd7b6a" (UID: "8aa711d1-ae51-480c-aa8b-9a510dfd7b6a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:03:16 crc kubenswrapper[4934]: I1227 08:03:16.916612 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "8aa711d1-ae51-480c-aa8b-9a510dfd7b6a" (UID: "8aa711d1-ae51-480c-aa8b-9a510dfd7b6a"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:03:16 crc kubenswrapper[4934]: I1227 08:03:16.920706 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-kube-api-access-cj5rv" (OuterVolumeSpecName: "kube-api-access-cj5rv") pod "8aa711d1-ae51-480c-aa8b-9a510dfd7b6a" (UID: "8aa711d1-ae51-480c-aa8b-9a510dfd7b6a"). InnerVolumeSpecName "kube-api-access-cj5rv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:16 crc kubenswrapper[4934]: I1227 08:03:16.923199 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "8aa711d1-ae51-480c-aa8b-9a510dfd7b6a" (UID: "8aa711d1-ae51-480c-aa8b-9a510dfd7b6a"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:03:16 crc kubenswrapper[4934]: I1227 08:03:16.923870 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "8aa711d1-ae51-480c-aa8b-9a510dfd7b6a" (UID: "8aa711d1-ae51-480c-aa8b-9a510dfd7b6a"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.017396 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cj5rv\" (UniqueName: \"kubernetes.io/projected/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-kube-api-access-cj5rv\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.017426 4934 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.017436 4934 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-service-ca\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.017445 4934 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.017453 4934 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.017464 4934 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.017473 4934 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a-console-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.046120 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.187256 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f752-account-create-mxw52"] Dec 27 08:03:17 crc kubenswrapper[4934]: W1227 08:03:17.192727 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod766e151c_8962_4bd5_b174_4e381730ff6f.slice/crio-f64748e034101247bb5855ae47054a017603af3df6563fd377192f96d1db95fc WatchSource:0}: Error finding container f64748e034101247bb5855ae47054a017603af3df6563fd377192f96d1db95fc: Status 404 returned error can't find the container with id f64748e034101247bb5855ae47054a017603af3df6563fd377192f96d1db95fc Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.222046 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca9c1959-3f6a-4c4d-ab56-b464625d92c8-dns-svc\") pod \"ca9c1959-3f6a-4c4d-ab56-b464625d92c8\" (UID: \"ca9c1959-3f6a-4c4d-ab56-b464625d92c8\") " Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.222272 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca9c1959-3f6a-4c4d-ab56-b464625d92c8-config\") pod \"ca9c1959-3f6a-4c4d-ab56-b464625d92c8\" (UID: \"ca9c1959-3f6a-4c4d-ab56-b464625d92c8\") " Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.222406 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-955sp\" (UniqueName: \"kubernetes.io/projected/ca9c1959-3f6a-4c4d-ab56-b464625d92c8-kube-api-access-955sp\") pod \"ca9c1959-3f6a-4c4d-ab56-b464625d92c8\" (UID: \"ca9c1959-3f6a-4c4d-ab56-b464625d92c8\") " Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.238804 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca9c1959-3f6a-4c4d-ab56-b464625d92c8-kube-api-access-955sp" (OuterVolumeSpecName: "kube-api-access-955sp") pod "ca9c1959-3f6a-4c4d-ab56-b464625d92c8" (UID: "ca9c1959-3f6a-4c4d-ab56-b464625d92c8"). InnerVolumeSpecName "kube-api-access-955sp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.292685 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca9c1959-3f6a-4c4d-ab56-b464625d92c8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ca9c1959-3f6a-4c4d-ab56-b464625d92c8" (UID: "ca9c1959-3f6a-4c4d-ab56-b464625d92c8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.317315 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca9c1959-3f6a-4c4d-ab56-b464625d92c8-config" (OuterVolumeSpecName: "config") pod "ca9c1959-3f6a-4c4d-ab56-b464625d92c8" (UID: "ca9c1959-3f6a-4c4d-ab56-b464625d92c8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.323124 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f9d8-account-create-dxnbx"] Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.324380 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca9c1959-3f6a-4c4d-ab56-b464625d92c8-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.324399 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-955sp\" (UniqueName: \"kubernetes.io/projected/ca9c1959-3f6a-4c4d-ab56-b464625d92c8-kube-api-access-955sp\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.324410 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ca9c1959-3f6a-4c4d-ab56-b464625d92c8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.332903 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-622b-account-create-rz2xx"] Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.430639 4934 generic.go:334] "Generic (PLEG): container finished" podID="7ede7d3e-b268-4132-b5d3-70569e23f060" containerID="c67d0c4914ff38adf00e96804612222913486319608a18a6c89859be0377e391" exitCode=0 Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.430732 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-cgffw" event={"ID":"7ede7d3e-b268-4132-b5d3-70569e23f060","Type":"ContainerDied","Data":"c67d0c4914ff38adf00e96804612222913486319608a18a6c89859be0377e391"} Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.437649 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" event={"ID":"ca9c1959-3f6a-4c4d-ab56-b464625d92c8","Type":"ContainerDied","Data":"cfcee839596b32cac9b5aeae2016f77d7ce8f5fe5896a08464822294fd807a87"} Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.437706 4934 scope.go:117] "RemoveContainer" containerID="8ef7f2c1ecadd3836be5e7f839be16b66848904a88d51eb92ee6d74ee50d8f77" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.437862 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5fzqq" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.443953 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-622b-account-create-rz2xx" event={"ID":"4e6e9101-e025-4566-833b-fd29fa46ab79","Type":"ContainerStarted","Data":"ded054edf86fb164dbe6b961675c58b5b5a8af63531058de69a22936aa01a02a"} Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.449675 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92d66221-653f-4837-a733-f02073b6ed70","Type":"ContainerStarted","Data":"12b822224b3e8dc94c80b53120e42ed082e0f0e3b3001ded509f8d22b739a9e4"} Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.454959 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"5788f6c6-2500-4c66-af7a-830cde17caa4","Type":"ContainerStarted","Data":"e635ca483a8f18a75caaaa57d2ae4d90d80def0795952c3144ec97d276df4322"} Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.455206 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-1" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.461697 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f752-account-create-mxw52" event={"ID":"766e151c-8962-4bd5-b174-4e381730ff6f","Type":"ContainerStarted","Data":"f64748e034101247bb5855ae47054a017603af3df6563fd377192f96d1db95fc"} Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.466054 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bswkq" event={"ID":"ccb12f0b-32d5-4ab4-9228-a8aa005b1304","Type":"ContainerStarted","Data":"d55ebc68748897501ff19659307be2f28c9bd6d2cf5e5369d4f5fda9394a7c54"} Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.482068 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"df7d6b06-cb87-4635-8aca-1ab00a564dd7","Type":"ContainerStarted","Data":"7572ade40e346d8651cbf1ce8fabe437ad6fbf893725ab3cc22288f68c146c30"} Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.482276 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-2" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.489620 4934 scope.go:117] "RemoveContainer" containerID="6cbabc9006c4fee872f9dd8e09e62ed1b75c67e72fe49d76d2fc6a2033c09151" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.490539 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f9d8-account-create-dxnbx" event={"ID":"84b8ef67-1fd1-4d7c-82d3-cde56ca1468a","Type":"ContainerStarted","Data":"2aa3d08193c8a3d3120244567a443bca71b1e51117feb5c5574e09c73e9fba53"} Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.509162 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5fzqq"] Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.509853 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-86b5d5c74c-c9lp9_8aa711d1-ae51-480c-aa8b-9a510dfd7b6a/console/0.log" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.509938 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-86b5d5c74c-c9lp9" event={"ID":"8aa711d1-ae51-480c-aa8b-9a510dfd7b6a","Type":"ContainerDied","Data":"32a5fad3b33b80f1a9941ce5b483e68c94c17a6b404184f3df8a23d351fecc9b"} Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.510006 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-86b5d5c74c-c9lp9" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.520963 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"76158d63-2099-4418-82a2-09af0432c1a3","Type":"ContainerStarted","Data":"d137c3d7f1fb61a5492ae4e87102e9b26b5caceeac796ed7e5db20c3505e981d"} Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.521785 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.534249 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c","Type":"ContainerStarted","Data":"00e3619755c94f9d80fd73245a68300e5f89f6b0603f9ba2025aa1302d112f7d"} Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.535158 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.536443 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5fzqq"] Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.545004 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-1" podStartSLOduration=41.409125485 podStartE2EDuration="1m0.544983004s" podCreationTimestamp="2025-12-27 08:02:17 +0000 UTC" firstStartedPulling="2025-12-27 08:02:19.468404055 +0000 UTC m=+1200.288844639" lastFinishedPulling="2025-12-27 08:02:38.604261554 +0000 UTC m=+1219.424702158" observedRunningTime="2025-12-27 08:03:17.526748356 +0000 UTC m=+1258.347188950" watchObservedRunningTime="2025-12-27 08:03:17.544983004 +0000 UTC m=+1258.365423598" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.567517 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-bswkq" podStartSLOduration=2.267206174 podStartE2EDuration="12.567492618s" podCreationTimestamp="2025-12-27 08:03:05 +0000 UTC" firstStartedPulling="2025-12-27 08:03:06.372573314 +0000 UTC m=+1247.193013918" lastFinishedPulling="2025-12-27 08:03:16.672859768 +0000 UTC m=+1257.493300362" observedRunningTime="2025-12-27 08:03:17.556438931 +0000 UTC m=+1258.376879545" watchObservedRunningTime="2025-12-27 08:03:17.567492618 +0000 UTC m=+1258.387933212" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.585051 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=41.449429325 podStartE2EDuration="1m0.585033188s" podCreationTimestamp="2025-12-27 08:02:17 +0000 UTC" firstStartedPulling="2025-12-27 08:02:19.589823551 +0000 UTC m=+1200.410264145" lastFinishedPulling="2025-12-27 08:02:38.725427424 +0000 UTC m=+1219.545868008" observedRunningTime="2025-12-27 08:03:17.580099764 +0000 UTC m=+1258.400540358" watchObservedRunningTime="2025-12-27 08:03:17.585033188 +0000 UTC m=+1258.405473782" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.626950 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=41.231435828 podStartE2EDuration="1m0.626931579s" podCreationTimestamp="2025-12-27 08:02:17 +0000 UTC" firstStartedPulling="2025-12-27 08:02:19.333312667 +0000 UTC m=+1200.153753261" lastFinishedPulling="2025-12-27 08:02:38.728808418 +0000 UTC m=+1219.549249012" observedRunningTime="2025-12-27 08:03:17.617674827 +0000 UTC m=+1258.438115421" watchObservedRunningTime="2025-12-27 08:03:17.626931579 +0000 UTC m=+1258.447372173" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.648887 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-2" podStartSLOduration=41.468796221 podStartE2EDuration="1m0.648871779s" podCreationTimestamp="2025-12-27 08:02:17 +0000 UTC" firstStartedPulling="2025-12-27 08:02:19.443916941 +0000 UTC m=+1200.264357535" lastFinishedPulling="2025-12-27 08:02:38.623992499 +0000 UTC m=+1219.444433093" observedRunningTime="2025-12-27 08:03:17.646635023 +0000 UTC m=+1258.467075617" watchObservedRunningTime="2025-12-27 08:03:17.648871779 +0000 UTC m=+1258.469312373" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.664147 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-86b5d5c74c-c9lp9"] Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.675295 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-86b5d5c74c-c9lp9"] Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.750410 4934 scope.go:117] "RemoveContainer" containerID="eacd15e000e9c4e9be81a96aad383398017a9643d62d695c37cfa5fb8b300d93" Dec 27 08:03:17 crc kubenswrapper[4934]: I1227 08:03:17.974586 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 27 08:03:18 crc kubenswrapper[4934]: I1227 08:03:18.547795 4934 generic.go:334] "Generic (PLEG): container finished" podID="4e6e9101-e025-4566-833b-fd29fa46ab79" containerID="9bdfc08d4542de65fa651c74626a5caf350ea01f0662434b7d89dc99e7088a05" exitCode=0 Dec 27 08:03:18 crc kubenswrapper[4934]: I1227 08:03:18.548237 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-622b-account-create-rz2xx" event={"ID":"4e6e9101-e025-4566-833b-fd29fa46ab79","Type":"ContainerDied","Data":"9bdfc08d4542de65fa651c74626a5caf350ea01f0662434b7d89dc99e7088a05"} Dec 27 08:03:18 crc kubenswrapper[4934]: I1227 08:03:18.553071 4934 generic.go:334] "Generic (PLEG): container finished" podID="766e151c-8962-4bd5-b174-4e381730ff6f" containerID="0249d45e628cf667116fcfea30516924fefff2ec7c5ae9b97447a5e0d5cf5011" exitCode=0 Dec 27 08:03:18 crc kubenswrapper[4934]: I1227 08:03:18.553152 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f752-account-create-mxw52" event={"ID":"766e151c-8962-4bd5-b174-4e381730ff6f","Type":"ContainerDied","Data":"0249d45e628cf667116fcfea30516924fefff2ec7c5ae9b97447a5e0d5cf5011"} Dec 27 08:03:18 crc kubenswrapper[4934]: I1227 08:03:18.554463 4934 generic.go:334] "Generic (PLEG): container finished" podID="84b8ef67-1fd1-4d7c-82d3-cde56ca1468a" containerID="93f68cbe460b9e15249c70d01c06cbd2bdd7a3da61f34a0b42945ed88fab9f8f" exitCode=0 Dec 27 08:03:18 crc kubenswrapper[4934]: I1227 08:03:18.555149 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f9d8-account-create-dxnbx" event={"ID":"84b8ef67-1fd1-4d7c-82d3-cde56ca1468a","Type":"ContainerDied","Data":"93f68cbe460b9e15249c70d01c06cbd2bdd7a3da61f34a0b42945ed88fab9f8f"} Dec 27 08:03:19 crc kubenswrapper[4934]: I1227 08:03:19.052619 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cgffw" Dec 27 08:03:19 crc kubenswrapper[4934]: I1227 08:03:19.058762 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knpzf\" (UniqueName: \"kubernetes.io/projected/7ede7d3e-b268-4132-b5d3-70569e23f060-kube-api-access-knpzf\") pod \"7ede7d3e-b268-4132-b5d3-70569e23f060\" (UID: \"7ede7d3e-b268-4132-b5d3-70569e23f060\") " Dec 27 08:03:19 crc kubenswrapper[4934]: I1227 08:03:19.068420 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ede7d3e-b268-4132-b5d3-70569e23f060-kube-api-access-knpzf" (OuterVolumeSpecName: "kube-api-access-knpzf") pod "7ede7d3e-b268-4132-b5d3-70569e23f060" (UID: "7ede7d3e-b268-4132-b5d3-70569e23f060"). InnerVolumeSpecName "kube-api-access-knpzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:19 crc kubenswrapper[4934]: I1227 08:03:19.159952 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knpzf\" (UniqueName: \"kubernetes.io/projected/7ede7d3e-b268-4132-b5d3-70569e23f060-kube-api-access-knpzf\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:19 crc kubenswrapper[4934]: I1227 08:03:19.478707 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8aa711d1-ae51-480c-aa8b-9a510dfd7b6a" path="/var/lib/kubelet/pods/8aa711d1-ae51-480c-aa8b-9a510dfd7b6a/volumes" Dec 27 08:03:19 crc kubenswrapper[4934]: I1227 08:03:19.479235 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca9c1959-3f6a-4c4d-ab56-b464625d92c8" path="/var/lib/kubelet/pods/ca9c1959-3f6a-4c4d-ab56-b464625d92c8/volumes" Dec 27 08:03:19 crc kubenswrapper[4934]: I1227 08:03:19.572911 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cgffw" Dec 27 08:03:19 crc kubenswrapper[4934]: I1227 08:03:19.573474 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-cgffw" event={"ID":"7ede7d3e-b268-4132-b5d3-70569e23f060","Type":"ContainerDied","Data":"72350978495f998b78cb3e951e8951247640c946c6e07cfd3933954fd342fb9e"} Dec 27 08:03:19 crc kubenswrapper[4934]: I1227 08:03:19.573612 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72350978495f998b78cb3e951e8951247640c946c6e07cfd3933954fd342fb9e" Dec 27 08:03:20 crc kubenswrapper[4934]: I1227 08:03:20.029262 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-622b-account-create-rz2xx" Dec 27 08:03:20 crc kubenswrapper[4934]: I1227 08:03:20.178235 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvlhp\" (UniqueName: \"kubernetes.io/projected/4e6e9101-e025-4566-833b-fd29fa46ab79-kube-api-access-nvlhp\") pod \"4e6e9101-e025-4566-833b-fd29fa46ab79\" (UID: \"4e6e9101-e025-4566-833b-fd29fa46ab79\") " Dec 27 08:03:20 crc kubenswrapper[4934]: I1227 08:03:20.184834 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e6e9101-e025-4566-833b-fd29fa46ab79-kube-api-access-nvlhp" (OuterVolumeSpecName: "kube-api-access-nvlhp") pod "4e6e9101-e025-4566-833b-fd29fa46ab79" (UID: "4e6e9101-e025-4566-833b-fd29fa46ab79"). InnerVolumeSpecName "kube-api-access-nvlhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:20 crc kubenswrapper[4934]: I1227 08:03:20.281838 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvlhp\" (UniqueName: \"kubernetes.io/projected/4e6e9101-e025-4566-833b-fd29fa46ab79-kube-api-access-nvlhp\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:20 crc kubenswrapper[4934]: I1227 08:03:20.583269 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f9d8-account-create-dxnbx" event={"ID":"84b8ef67-1fd1-4d7c-82d3-cde56ca1468a","Type":"ContainerDied","Data":"2aa3d08193c8a3d3120244567a443bca71b1e51117feb5c5574e09c73e9fba53"} Dec 27 08:03:20 crc kubenswrapper[4934]: I1227 08:03:20.583307 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2aa3d08193c8a3d3120244567a443bca71b1e51117feb5c5574e09c73e9fba53" Dec 27 08:03:20 crc kubenswrapper[4934]: I1227 08:03:20.584817 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f752-account-create-mxw52" event={"ID":"766e151c-8962-4bd5-b174-4e381730ff6f","Type":"ContainerDied","Data":"f64748e034101247bb5855ae47054a017603af3df6563fd377192f96d1db95fc"} Dec 27 08:03:20 crc kubenswrapper[4934]: I1227 08:03:20.584862 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f64748e034101247bb5855ae47054a017603af3df6563fd377192f96d1db95fc" Dec 27 08:03:20 crc kubenswrapper[4934]: I1227 08:03:20.586190 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-622b-account-create-rz2xx" event={"ID":"4e6e9101-e025-4566-833b-fd29fa46ab79","Type":"ContainerDied","Data":"ded054edf86fb164dbe6b961675c58b5b5a8af63531058de69a22936aa01a02a"} Dec 27 08:03:20 crc kubenswrapper[4934]: I1227 08:03:20.586216 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ded054edf86fb164dbe6b961675c58b5b5a8af63531058de69a22936aa01a02a" Dec 27 08:03:20 crc kubenswrapper[4934]: I1227 08:03:20.586277 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-622b-account-create-rz2xx" Dec 27 08:03:20 crc kubenswrapper[4934]: I1227 08:03:20.728105 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f752-account-create-mxw52" Dec 27 08:03:20 crc kubenswrapper[4934]: I1227 08:03:20.735097 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f9d8-account-create-dxnbx" Dec 27 08:03:20 crc kubenswrapper[4934]: I1227 08:03:20.802298 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-etc-swift\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") " pod="openstack/swift-storage-0" Dec 27 08:03:20 crc kubenswrapper[4934]: E1227 08:03:20.802544 4934 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 27 08:03:20 crc kubenswrapper[4934]: E1227 08:03:20.802583 4934 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 27 08:03:20 crc kubenswrapper[4934]: E1227 08:03:20.802657 4934 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-etc-swift podName:e59489eb-aaa1-4188-b593-d3d3c2f1f412 nodeName:}" failed. No retries permitted until 2025-12-27 08:03:36.802634906 +0000 UTC m=+1277.623075510 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-etc-swift") pod "swift-storage-0" (UID: "e59489eb-aaa1-4188-b593-d3d3c2f1f412") : configmap "swift-ring-files" not found Dec 27 08:03:20 crc kubenswrapper[4934]: I1227 08:03:20.903586 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnrnd\" (UniqueName: \"kubernetes.io/projected/766e151c-8962-4bd5-b174-4e381730ff6f-kube-api-access-dnrnd\") pod \"766e151c-8962-4bd5-b174-4e381730ff6f\" (UID: \"766e151c-8962-4bd5-b174-4e381730ff6f\") " Dec 27 08:03:20 crc kubenswrapper[4934]: I1227 08:03:20.903633 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l99th\" (UniqueName: \"kubernetes.io/projected/84b8ef67-1fd1-4d7c-82d3-cde56ca1468a-kube-api-access-l99th\") pod \"84b8ef67-1fd1-4d7c-82d3-cde56ca1468a\" (UID: \"84b8ef67-1fd1-4d7c-82d3-cde56ca1468a\") " Dec 27 08:03:20 crc kubenswrapper[4934]: I1227 08:03:20.939198 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/766e151c-8962-4bd5-b174-4e381730ff6f-kube-api-access-dnrnd" (OuterVolumeSpecName: "kube-api-access-dnrnd") pod "766e151c-8962-4bd5-b174-4e381730ff6f" (UID: "766e151c-8962-4bd5-b174-4e381730ff6f"). InnerVolumeSpecName "kube-api-access-dnrnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:20 crc kubenswrapper[4934]: I1227 08:03:20.939325 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84b8ef67-1fd1-4d7c-82d3-cde56ca1468a-kube-api-access-l99th" (OuterVolumeSpecName: "kube-api-access-l99th") pod "84b8ef67-1fd1-4d7c-82d3-cde56ca1468a" (UID: "84b8ef67-1fd1-4d7c-82d3-cde56ca1468a"). InnerVolumeSpecName "kube-api-access-l99th". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:21 crc kubenswrapper[4934]: I1227 08:03:21.006013 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnrnd\" (UniqueName: \"kubernetes.io/projected/766e151c-8962-4bd5-b174-4e381730ff6f-kube-api-access-dnrnd\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:21 crc kubenswrapper[4934]: I1227 08:03:21.006052 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l99th\" (UniqueName: \"kubernetes.io/projected/84b8ef67-1fd1-4d7c-82d3-cde56ca1468a-kube-api-access-l99th\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:21 crc kubenswrapper[4934]: I1227 08:03:21.598366 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92d66221-653f-4837-a733-f02073b6ed70","Type":"ContainerStarted","Data":"46acee5837df489cb9e18ac9daa2f27e318770779025b31476deffa809d0a809"} Dec 27 08:03:21 crc kubenswrapper[4934]: I1227 08:03:21.598399 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f9d8-account-create-dxnbx" Dec 27 08:03:21 crc kubenswrapper[4934]: I1227 08:03:21.598420 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f752-account-create-mxw52" Dec 27 08:03:22 crc kubenswrapper[4934]: I1227 08:03:22.782621 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-22dwz" podUID="f4b9eef2-987b-4c54-803c-6655aceab8f6" containerName="ovn-controller" probeResult="failure" output=< Dec 27 08:03:22 crc kubenswrapper[4934]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 27 08:03:22 crc kubenswrapper[4934]: > Dec 27 08:03:24 crc kubenswrapper[4934]: I1227 08:03:24.094327 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-mbcsv"] Dec 27 08:03:24 crc kubenswrapper[4934]: E1227 08:03:24.094946 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e6e9101-e025-4566-833b-fd29fa46ab79" containerName="mariadb-account-create" Dec 27 08:03:24 crc kubenswrapper[4934]: I1227 08:03:24.094957 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e6e9101-e025-4566-833b-fd29fa46ab79" containerName="mariadb-account-create" Dec 27 08:03:24 crc kubenswrapper[4934]: E1227 08:03:24.094976 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aa711d1-ae51-480c-aa8b-9a510dfd7b6a" containerName="console" Dec 27 08:03:24 crc kubenswrapper[4934]: I1227 08:03:24.094982 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aa711d1-ae51-480c-aa8b-9a510dfd7b6a" containerName="console" Dec 27 08:03:24 crc kubenswrapper[4934]: E1227 08:03:24.094999 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca9c1959-3f6a-4c4d-ab56-b464625d92c8" containerName="dnsmasq-dns" Dec 27 08:03:24 crc kubenswrapper[4934]: I1227 08:03:24.095005 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca9c1959-3f6a-4c4d-ab56-b464625d92c8" containerName="dnsmasq-dns" Dec 27 08:03:24 crc kubenswrapper[4934]: E1227 08:03:24.095015 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="766e151c-8962-4bd5-b174-4e381730ff6f" containerName="mariadb-account-create" Dec 27 08:03:24 crc kubenswrapper[4934]: I1227 08:03:24.095021 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="766e151c-8962-4bd5-b174-4e381730ff6f" containerName="mariadb-account-create" Dec 27 08:03:24 crc kubenswrapper[4934]: E1227 08:03:24.095043 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca9c1959-3f6a-4c4d-ab56-b464625d92c8" containerName="init" Dec 27 08:03:24 crc kubenswrapper[4934]: I1227 08:03:24.095048 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca9c1959-3f6a-4c4d-ab56-b464625d92c8" containerName="init" Dec 27 08:03:24 crc kubenswrapper[4934]: E1227 08:03:24.095058 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84b8ef67-1fd1-4d7c-82d3-cde56ca1468a" containerName="mariadb-account-create" Dec 27 08:03:24 crc kubenswrapper[4934]: I1227 08:03:24.095064 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="84b8ef67-1fd1-4d7c-82d3-cde56ca1468a" containerName="mariadb-account-create" Dec 27 08:03:24 crc kubenswrapper[4934]: E1227 08:03:24.095095 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ede7d3e-b268-4132-b5d3-70569e23f060" containerName="mariadb-database-create" Dec 27 08:03:24 crc kubenswrapper[4934]: I1227 08:03:24.095101 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ede7d3e-b268-4132-b5d3-70569e23f060" containerName="mariadb-database-create" Dec 27 08:03:24 crc kubenswrapper[4934]: I1227 08:03:24.095268 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ede7d3e-b268-4132-b5d3-70569e23f060" containerName="mariadb-database-create" Dec 27 08:03:24 crc kubenswrapper[4934]: I1227 08:03:24.095283 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca9c1959-3f6a-4c4d-ab56-b464625d92c8" containerName="dnsmasq-dns" Dec 27 08:03:24 crc kubenswrapper[4934]: I1227 08:03:24.095298 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8aa711d1-ae51-480c-aa8b-9a510dfd7b6a" containerName="console" Dec 27 08:03:24 crc kubenswrapper[4934]: I1227 08:03:24.095307 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="766e151c-8962-4bd5-b174-4e381730ff6f" containerName="mariadb-account-create" Dec 27 08:03:24 crc kubenswrapper[4934]: I1227 08:03:24.095318 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="84b8ef67-1fd1-4d7c-82d3-cde56ca1468a" containerName="mariadb-account-create" Dec 27 08:03:24 crc kubenswrapper[4934]: I1227 08:03:24.095330 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e6e9101-e025-4566-833b-fd29fa46ab79" containerName="mariadb-account-create" Dec 27 08:03:24 crc kubenswrapper[4934]: I1227 08:03:24.096005 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-mbcsv" Dec 27 08:03:24 crc kubenswrapper[4934]: I1227 08:03:24.113171 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-mbcsv"] Dec 27 08:03:24 crc kubenswrapper[4934]: I1227 08:03:24.173966 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5lt4\" (UniqueName: \"kubernetes.io/projected/59b9ab13-2205-48a4-8a33-9d85f7de7d04-kube-api-access-j5lt4\") pod \"mysqld-exporter-openstack-cell1-db-create-mbcsv\" (UID: \"59b9ab13-2205-48a4-8a33-9d85f7de7d04\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-mbcsv" Dec 27 08:03:24 crc kubenswrapper[4934]: I1227 08:03:24.275598 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5lt4\" (UniqueName: \"kubernetes.io/projected/59b9ab13-2205-48a4-8a33-9d85f7de7d04-kube-api-access-j5lt4\") pod \"mysqld-exporter-openstack-cell1-db-create-mbcsv\" (UID: \"59b9ab13-2205-48a4-8a33-9d85f7de7d04\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-mbcsv" Dec 27 08:03:24 crc kubenswrapper[4934]: I1227 08:03:24.384746 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5lt4\" (UniqueName: \"kubernetes.io/projected/59b9ab13-2205-48a4-8a33-9d85f7de7d04-kube-api-access-j5lt4\") pod \"mysqld-exporter-openstack-cell1-db-create-mbcsv\" (UID: \"59b9ab13-2205-48a4-8a33-9d85f7de7d04\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-mbcsv" Dec 27 08:03:24 crc kubenswrapper[4934]: I1227 08:03:24.415919 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-mbcsv" Dec 27 08:03:25 crc kubenswrapper[4934]: I1227 08:03:25.032114 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-mbcsv"] Dec 27 08:03:25 crc kubenswrapper[4934]: I1227 08:03:25.691444 4934 generic.go:334] "Generic (PLEG): container finished" podID="ccb12f0b-32d5-4ab4-9228-a8aa005b1304" containerID="d55ebc68748897501ff19659307be2f28c9bd6d2cf5e5369d4f5fda9394a7c54" exitCode=0 Dec 27 08:03:25 crc kubenswrapper[4934]: I1227 08:03:25.691530 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bswkq" event={"ID":"ccb12f0b-32d5-4ab4-9228-a8aa005b1304","Type":"ContainerDied","Data":"d55ebc68748897501ff19659307be2f28c9bd6d2cf5e5369d4f5fda9394a7c54"} Dec 27 08:03:25 crc kubenswrapper[4934]: I1227 08:03:25.694390 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92d66221-653f-4837-a733-f02073b6ed70","Type":"ContainerStarted","Data":"80a5156434ae03cae6f2c876737e9a1f41e83e3a35630adcd0abe3e0277e0270"} Dec 27 08:03:25 crc kubenswrapper[4934]: I1227 08:03:25.696031 4934 generic.go:334] "Generic (PLEG): container finished" podID="59b9ab13-2205-48a4-8a33-9d85f7de7d04" containerID="96d7f9fd2a893d377ce92367a2ddadd2206dbd691c9c4f16919c7d3017c33831" exitCode=0 Dec 27 08:03:25 crc kubenswrapper[4934]: I1227 08:03:25.696147 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-mbcsv" event={"ID":"59b9ab13-2205-48a4-8a33-9d85f7de7d04","Type":"ContainerDied","Data":"96d7f9fd2a893d377ce92367a2ddadd2206dbd691c9c4f16919c7d3017c33831"} Dec 27 08:03:25 crc kubenswrapper[4934]: I1227 08:03:25.696185 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-mbcsv" event={"ID":"59b9ab13-2205-48a4-8a33-9d85f7de7d04","Type":"ContainerStarted","Data":"8a2c88f913fa0c725ef44feb90969b3c362284935133046ee76d088b5e8010c1"} Dec 27 08:03:25 crc kubenswrapper[4934]: I1227 08:03:25.746375 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=17.49181256 podStartE2EDuration="1m2.74635155s" podCreationTimestamp="2025-12-27 08:02:23 +0000 UTC" firstStartedPulling="2025-12-27 08:02:39.268886855 +0000 UTC m=+1220.089327449" lastFinishedPulling="2025-12-27 08:03:24.523425845 +0000 UTC m=+1265.343866439" observedRunningTime="2025-12-27 08:03:25.738005071 +0000 UTC m=+1266.558445665" watchObservedRunningTime="2025-12-27 08:03:25.74635155 +0000 UTC m=+1266.566792164" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.072218 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-090b-account-create-89jvs"] Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.074298 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-090b-account-create-89jvs" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.077750 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.088737 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-090b-account-create-89jvs"] Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.239275 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8x6c\" (UniqueName: \"kubernetes.io/projected/9bb54f3d-f01e-445f-beee-2a1f33537008-kube-api-access-z8x6c\") pod \"glance-090b-account-create-89jvs\" (UID: \"9bb54f3d-f01e-445f-beee-2a1f33537008\") " pod="openstack/glance-090b-account-create-89jvs" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.313516 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.318954 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-mbcsv" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.341283 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8x6c\" (UniqueName: \"kubernetes.io/projected/9bb54f3d-f01e-445f-beee-2a1f33537008-kube-api-access-z8x6c\") pod \"glance-090b-account-create-89jvs\" (UID: \"9bb54f3d-f01e-445f-beee-2a1f33537008\") " pod="openstack/glance-090b-account-create-89jvs" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.361215 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8x6c\" (UniqueName: \"kubernetes.io/projected/9bb54f3d-f01e-445f-beee-2a1f33537008-kube-api-access-z8x6c\") pod \"glance-090b-account-create-89jvs\" (UID: \"9bb54f3d-f01e-445f-beee-2a1f33537008\") " pod="openstack/glance-090b-account-create-89jvs" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.400489 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-090b-account-create-89jvs" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.443191 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tv65\" (UniqueName: \"kubernetes.io/projected/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-kube-api-access-6tv65\") pod \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.443279 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5lt4\" (UniqueName: \"kubernetes.io/projected/59b9ab13-2205-48a4-8a33-9d85f7de7d04-kube-api-access-j5lt4\") pod \"59b9ab13-2205-48a4-8a33-9d85f7de7d04\" (UID: \"59b9ab13-2205-48a4-8a33-9d85f7de7d04\") " Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.443314 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-combined-ca-bundle\") pod \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.443424 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-swiftconf\") pod \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.443486 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-ring-data-devices\") pod \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.443525 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-dispersionconf\") pod \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.443581 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-scripts\") pod \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.443622 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-etc-swift\") pod \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\" (UID: \"ccb12f0b-32d5-4ab4-9228-a8aa005b1304\") " Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.444184 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "ccb12f0b-32d5-4ab4-9228-a8aa005b1304" (UID: "ccb12f0b-32d5-4ab4-9228-a8aa005b1304"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.444752 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ccb12f0b-32d5-4ab4-9228-a8aa005b1304" (UID: "ccb12f0b-32d5-4ab4-9228-a8aa005b1304"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.449443 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59b9ab13-2205-48a4-8a33-9d85f7de7d04-kube-api-access-j5lt4" (OuterVolumeSpecName: "kube-api-access-j5lt4") pod "59b9ab13-2205-48a4-8a33-9d85f7de7d04" (UID: "59b9ab13-2205-48a4-8a33-9d85f7de7d04"). InnerVolumeSpecName "kube-api-access-j5lt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.463233 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-kube-api-access-6tv65" (OuterVolumeSpecName: "kube-api-access-6tv65") pod "ccb12f0b-32d5-4ab4-9228-a8aa005b1304" (UID: "ccb12f0b-32d5-4ab4-9228-a8aa005b1304"). InnerVolumeSpecName "kube-api-access-6tv65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.470199 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "ccb12f0b-32d5-4ab4-9228-a8aa005b1304" (UID: "ccb12f0b-32d5-4ab4-9228-a8aa005b1304"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.492765 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ccb12f0b-32d5-4ab4-9228-a8aa005b1304" (UID: "ccb12f0b-32d5-4ab4-9228-a8aa005b1304"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.498219 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-scripts" (OuterVolumeSpecName: "scripts") pod "ccb12f0b-32d5-4ab4-9228-a8aa005b1304" (UID: "ccb12f0b-32d5-4ab4-9228-a8aa005b1304"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.505459 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "ccb12f0b-32d5-4ab4-9228-a8aa005b1304" (UID: "ccb12f0b-32d5-4ab4-9228-a8aa005b1304"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.546187 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.546209 4934 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.546221 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tv65\" (UniqueName: \"kubernetes.io/projected/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-kube-api-access-6tv65\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.546233 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5lt4\" (UniqueName: \"kubernetes.io/projected/59b9ab13-2205-48a4-8a33-9d85f7de7d04-kube-api-access-j5lt4\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.546244 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.546251 4934 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.546259 4934 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.546268 4934 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/ccb12f0b-32d5-4ab4-9228-a8aa005b1304-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.714452 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-bswkq" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.714472 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-bswkq" event={"ID":"ccb12f0b-32d5-4ab4-9228-a8aa005b1304","Type":"ContainerDied","Data":"7807c4891b8905cf95b8e80bc4e9874aeb5d16aea8fb3de2b4defa363d31f09f"} Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.714512 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7807c4891b8905cf95b8e80bc4e9874aeb5d16aea8fb3de2b4defa363d31f09f" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.716699 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-mbcsv" event={"ID":"59b9ab13-2205-48a4-8a33-9d85f7de7d04","Type":"ContainerDied","Data":"8a2c88f913fa0c725ef44feb90969b3c362284935133046ee76d088b5e8010c1"} Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.716744 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a2c88f913fa0c725ef44feb90969b3c362284935133046ee76d088b5e8010c1" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.716746 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-mbcsv" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.783817 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-22dwz" podUID="f4b9eef2-987b-4c54-803c-6655aceab8f6" containerName="ovn-controller" probeResult="failure" output=< Dec 27 08:03:27 crc kubenswrapper[4934]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 27 08:03:27 crc kubenswrapper[4934]: > Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.801271 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.804738 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-bdjjv" Dec 27 08:03:27 crc kubenswrapper[4934]: I1227 08:03:27.923365 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-090b-account-create-89jvs"] Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.036640 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-22dwz-config-p5hgz"] Dec 27 08:03:28 crc kubenswrapper[4934]: E1227 08:03:28.037049 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59b9ab13-2205-48a4-8a33-9d85f7de7d04" containerName="mariadb-database-create" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.037067 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="59b9ab13-2205-48a4-8a33-9d85f7de7d04" containerName="mariadb-database-create" Dec 27 08:03:28 crc kubenswrapper[4934]: E1227 08:03:28.037107 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccb12f0b-32d5-4ab4-9228-a8aa005b1304" containerName="swift-ring-rebalance" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.037114 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccb12f0b-32d5-4ab4-9228-a8aa005b1304" containerName="swift-ring-rebalance" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.037333 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccb12f0b-32d5-4ab4-9228-a8aa005b1304" containerName="swift-ring-rebalance" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.037357 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="59b9ab13-2205-48a4-8a33-9d85f7de7d04" containerName="mariadb-database-create" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.038016 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.040654 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.064892 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-22dwz-config-p5hgz"] Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.174017 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8a7c7644-3523-4533-8e7a-39652454a9a0-var-log-ovn\") pod \"ovn-controller-22dwz-config-p5hgz\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.175091 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8a7c7644-3523-4533-8e7a-39652454a9a0-additional-scripts\") pod \"ovn-controller-22dwz-config-p5hgz\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.175271 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a7c7644-3523-4533-8e7a-39652454a9a0-scripts\") pod \"ovn-controller-22dwz-config-p5hgz\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.175376 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffrfd\" (UniqueName: \"kubernetes.io/projected/8a7c7644-3523-4533-8e7a-39652454a9a0-kube-api-access-ffrfd\") pod \"ovn-controller-22dwz-config-p5hgz\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.175649 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8a7c7644-3523-4533-8e7a-39652454a9a0-var-run-ovn\") pod \"ovn-controller-22dwz-config-p5hgz\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.175731 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8a7c7644-3523-4533-8e7a-39652454a9a0-var-run\") pod \"ovn-controller-22dwz-config-p5hgz\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.277848 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a7c7644-3523-4533-8e7a-39652454a9a0-scripts\") pod \"ovn-controller-22dwz-config-p5hgz\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.278106 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffrfd\" (UniqueName: \"kubernetes.io/projected/8a7c7644-3523-4533-8e7a-39652454a9a0-kube-api-access-ffrfd\") pod \"ovn-controller-22dwz-config-p5hgz\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.278166 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8a7c7644-3523-4533-8e7a-39652454a9a0-var-run-ovn\") pod \"ovn-controller-22dwz-config-p5hgz\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.278192 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8a7c7644-3523-4533-8e7a-39652454a9a0-var-run\") pod \"ovn-controller-22dwz-config-p5hgz\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.278257 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8a7c7644-3523-4533-8e7a-39652454a9a0-var-log-ovn\") pod \"ovn-controller-22dwz-config-p5hgz\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.278279 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8a7c7644-3523-4533-8e7a-39652454a9a0-additional-scripts\") pod \"ovn-controller-22dwz-config-p5hgz\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.278756 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8a7c7644-3523-4533-8e7a-39652454a9a0-var-run\") pod \"ovn-controller-22dwz-config-p5hgz\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.278795 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8a7c7644-3523-4533-8e7a-39652454a9a0-var-log-ovn\") pod \"ovn-controller-22dwz-config-p5hgz\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.278915 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8a7c7644-3523-4533-8e7a-39652454a9a0-additional-scripts\") pod \"ovn-controller-22dwz-config-p5hgz\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.278907 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8a7c7644-3523-4533-8e7a-39652454a9a0-var-run-ovn\") pod \"ovn-controller-22dwz-config-p5hgz\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.279798 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a7c7644-3523-4533-8e7a-39652454a9a0-scripts\") pod \"ovn-controller-22dwz-config-p5hgz\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.295847 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffrfd\" (UniqueName: \"kubernetes.io/projected/8a7c7644-3523-4533-8e7a-39652454a9a0-kube-api-access-ffrfd\") pod \"ovn-controller-22dwz-config-p5hgz\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.353192 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.730726 4934 generic.go:334] "Generic (PLEG): container finished" podID="9bb54f3d-f01e-445f-beee-2a1f33537008" containerID="3f946b91baf02cf6d91daa86547ddb8ec839e63d7906f28412e858239ccaec2d" exitCode=0 Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.730800 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-090b-account-create-89jvs" event={"ID":"9bb54f3d-f01e-445f-beee-2a1f33537008","Type":"ContainerDied","Data":"3f946b91baf02cf6d91daa86547ddb8ec839e63d7906f28412e858239ccaec2d"} Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.731064 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-090b-account-create-89jvs" event={"ID":"9bb54f3d-f01e-445f-beee-2a1f33537008","Type":"ContainerStarted","Data":"3e6b569507f3c2a7e4a480bedbbfc7e1956e0058c79ea8777e6f619dce1edb23"} Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.812044 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.129:5671: connect: connection refused" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.832188 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-1" podUID="5788f6c6-2500-4c66-af7a-830cde17caa4" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.130:5671: connect: connection refused" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.851070 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="df7d6b06-cb87-4635-8aca-1ab00a564dd7" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.131:5671: connect: connection refused" Dec 27 08:03:28 crc kubenswrapper[4934]: I1227 08:03:28.871848 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-22dwz-config-p5hgz"] Dec 27 08:03:29 crc kubenswrapper[4934]: I1227 08:03:29.047338 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:03:29 crc kubenswrapper[4934]: I1227 08:03:29.741401 4934 generic.go:334] "Generic (PLEG): container finished" podID="8a7c7644-3523-4533-8e7a-39652454a9a0" containerID="aac373063ba829ef3ed599024d2a013dd99da808d9e914e732e6fc9b32afa8fd" exitCode=0 Dec 27 08:03:29 crc kubenswrapper[4934]: I1227 08:03:29.741488 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-22dwz-config-p5hgz" event={"ID":"8a7c7644-3523-4533-8e7a-39652454a9a0","Type":"ContainerDied","Data":"aac373063ba829ef3ed599024d2a013dd99da808d9e914e732e6fc9b32afa8fd"} Dec 27 08:03:29 crc kubenswrapper[4934]: I1227 08:03:29.742205 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-22dwz-config-p5hgz" event={"ID":"8a7c7644-3523-4533-8e7a-39652454a9a0","Type":"ContainerStarted","Data":"a50c9a0710e63ae885298c9ac59416e8f4e10713c673b20c033dbdd10e4399fe"} Dec 27 08:03:30 crc kubenswrapper[4934]: I1227 08:03:30.158409 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-090b-account-create-89jvs" Dec 27 08:03:30 crc kubenswrapper[4934]: I1227 08:03:30.272618 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:30 crc kubenswrapper[4934]: I1227 08:03:30.348765 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8x6c\" (UniqueName: \"kubernetes.io/projected/9bb54f3d-f01e-445f-beee-2a1f33537008-kube-api-access-z8x6c\") pod \"9bb54f3d-f01e-445f-beee-2a1f33537008\" (UID: \"9bb54f3d-f01e-445f-beee-2a1f33537008\") " Dec 27 08:03:30 crc kubenswrapper[4934]: I1227 08:03:30.360872 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bb54f3d-f01e-445f-beee-2a1f33537008-kube-api-access-z8x6c" (OuterVolumeSpecName: "kube-api-access-z8x6c") pod "9bb54f3d-f01e-445f-beee-2a1f33537008" (UID: "9bb54f3d-f01e-445f-beee-2a1f33537008"). InnerVolumeSpecName "kube-api-access-z8x6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:30 crc kubenswrapper[4934]: I1227 08:03:30.452874 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8x6c\" (UniqueName: \"kubernetes.io/projected/9bb54f3d-f01e-445f-beee-2a1f33537008-kube-api-access-z8x6c\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:30 crc kubenswrapper[4934]: I1227 08:03:30.756397 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-090b-account-create-89jvs" event={"ID":"9bb54f3d-f01e-445f-beee-2a1f33537008","Type":"ContainerDied","Data":"3e6b569507f3c2a7e4a480bedbbfc7e1956e0058c79ea8777e6f619dce1edb23"} Dec 27 08:03:30 crc kubenswrapper[4934]: I1227 08:03:30.756751 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e6b569507f3c2a7e4a480bedbbfc7e1956e0058c79ea8777e6f619dce1edb23" Dec 27 08:03:30 crc kubenswrapper[4934]: I1227 08:03:30.761240 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-090b-account-create-89jvs" Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.157681 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.272201 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8a7c7644-3523-4533-8e7a-39652454a9a0-var-run-ovn\") pod \"8a7c7644-3523-4533-8e7a-39652454a9a0\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.272362 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8a7c7644-3523-4533-8e7a-39652454a9a0-var-run\") pod \"8a7c7644-3523-4533-8e7a-39652454a9a0\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.272401 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffrfd\" (UniqueName: \"kubernetes.io/projected/8a7c7644-3523-4533-8e7a-39652454a9a0-kube-api-access-ffrfd\") pod \"8a7c7644-3523-4533-8e7a-39652454a9a0\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.272444 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8a7c7644-3523-4533-8e7a-39652454a9a0-var-log-ovn\") pod \"8a7c7644-3523-4533-8e7a-39652454a9a0\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.272511 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8a7c7644-3523-4533-8e7a-39652454a9a0-additional-scripts\") pod \"8a7c7644-3523-4533-8e7a-39652454a9a0\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.272609 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a7c7644-3523-4533-8e7a-39652454a9a0-scripts\") pod \"8a7c7644-3523-4533-8e7a-39652454a9a0\" (UID: \"8a7c7644-3523-4533-8e7a-39652454a9a0\") " Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.273280 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8a7c7644-3523-4533-8e7a-39652454a9a0-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "8a7c7644-3523-4533-8e7a-39652454a9a0" (UID: "8a7c7644-3523-4533-8e7a-39652454a9a0"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.273345 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8a7c7644-3523-4533-8e7a-39652454a9a0-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "8a7c7644-3523-4533-8e7a-39652454a9a0" (UID: "8a7c7644-3523-4533-8e7a-39652454a9a0"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.273398 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8a7c7644-3523-4533-8e7a-39652454a9a0-var-run" (OuterVolumeSpecName: "var-run") pod "8a7c7644-3523-4533-8e7a-39652454a9a0" (UID: "8a7c7644-3523-4533-8e7a-39652454a9a0"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.273963 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a7c7644-3523-4533-8e7a-39652454a9a0-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "8a7c7644-3523-4533-8e7a-39652454a9a0" (UID: "8a7c7644-3523-4533-8e7a-39652454a9a0"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.274602 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a7c7644-3523-4533-8e7a-39652454a9a0-scripts" (OuterVolumeSpecName: "scripts") pod "8a7c7644-3523-4533-8e7a-39652454a9a0" (UID: "8a7c7644-3523-4533-8e7a-39652454a9a0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.281813 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a7c7644-3523-4533-8e7a-39652454a9a0-kube-api-access-ffrfd" (OuterVolumeSpecName: "kube-api-access-ffrfd") pod "8a7c7644-3523-4533-8e7a-39652454a9a0" (UID: "8a7c7644-3523-4533-8e7a-39652454a9a0"). InnerVolumeSpecName "kube-api-access-ffrfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.374861 4934 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8a7c7644-3523-4533-8e7a-39652454a9a0-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.374895 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8a7c7644-3523-4533-8e7a-39652454a9a0-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.374903 4934 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8a7c7644-3523-4533-8e7a-39652454a9a0-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.374911 4934 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8a7c7644-3523-4533-8e7a-39652454a9a0-var-run\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.374919 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffrfd\" (UniqueName: \"kubernetes.io/projected/8a7c7644-3523-4533-8e7a-39652454a9a0-kube-api-access-ffrfd\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.374928 4934 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8a7c7644-3523-4533-8e7a-39652454a9a0-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.771297 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-22dwz-config-p5hgz" event={"ID":"8a7c7644-3523-4533-8e7a-39652454a9a0","Type":"ContainerDied","Data":"a50c9a0710e63ae885298c9ac59416e8f4e10713c673b20c033dbdd10e4399fe"} Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.771346 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a50c9a0710e63ae885298c9ac59416e8f4e10713c673b20c033dbdd10e4399fe" Dec 27 08:03:31 crc kubenswrapper[4934]: I1227 08:03:31.771404 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-22dwz-config-p5hgz" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.238630 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-6vlxg"] Dec 27 08:03:32 crc kubenswrapper[4934]: E1227 08:03:32.239138 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bb54f3d-f01e-445f-beee-2a1f33537008" containerName="mariadb-account-create" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.239161 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bb54f3d-f01e-445f-beee-2a1f33537008" containerName="mariadb-account-create" Dec 27 08:03:32 crc kubenswrapper[4934]: E1227 08:03:32.239175 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a7c7644-3523-4533-8e7a-39652454a9a0" containerName="ovn-config" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.239184 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a7c7644-3523-4533-8e7a-39652454a9a0" containerName="ovn-config" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.239432 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bb54f3d-f01e-445f-beee-2a1f33537008" containerName="mariadb-account-create" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.239476 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a7c7644-3523-4533-8e7a-39652454a9a0" containerName="ovn-config" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.240278 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6vlxg" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.244104 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-9x8k8" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.245002 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.252719 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-6vlxg"] Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.324328 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-22dwz-config-p5hgz"] Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.332345 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-22dwz-config-p5hgz"] Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.375286 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-22dwz-config-6s9mt"] Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.376670 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.384803 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.391144 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-22dwz-config-6s9mt"] Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.403281 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94526b19-793f-4ec9-9a80-8a7d4f8cab87-config-data\") pod \"glance-db-sync-6vlxg\" (UID: \"94526b19-793f-4ec9-9a80-8a7d4f8cab87\") " pod="openstack/glance-db-sync-6vlxg" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.403506 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94526b19-793f-4ec9-9a80-8a7d4f8cab87-combined-ca-bundle\") pod \"glance-db-sync-6vlxg\" (UID: \"94526b19-793f-4ec9-9a80-8a7d4f8cab87\") " pod="openstack/glance-db-sync-6vlxg" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.403537 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/94526b19-793f-4ec9-9a80-8a7d4f8cab87-db-sync-config-data\") pod \"glance-db-sync-6vlxg\" (UID: \"94526b19-793f-4ec9-9a80-8a7d4f8cab87\") " pod="openstack/glance-db-sync-6vlxg" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.403831 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt5pp\" (UniqueName: \"kubernetes.io/projected/94526b19-793f-4ec9-9a80-8a7d4f8cab87-kube-api-access-rt5pp\") pod \"glance-db-sync-6vlxg\" (UID: \"94526b19-793f-4ec9-9a80-8a7d4f8cab87\") " pod="openstack/glance-db-sync-6vlxg" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.505139 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szv5t\" (UniqueName: \"kubernetes.io/projected/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-kube-api-access-szv5t\") pod \"ovn-controller-22dwz-config-6s9mt\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.505276 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-var-run\") pod \"ovn-controller-22dwz-config-6s9mt\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.505314 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-var-log-ovn\") pod \"ovn-controller-22dwz-config-6s9mt\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.505339 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt5pp\" (UniqueName: \"kubernetes.io/projected/94526b19-793f-4ec9-9a80-8a7d4f8cab87-kube-api-access-rt5pp\") pod \"glance-db-sync-6vlxg\" (UID: \"94526b19-793f-4ec9-9a80-8a7d4f8cab87\") " pod="openstack/glance-db-sync-6vlxg" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.505360 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-additional-scripts\") pod \"ovn-controller-22dwz-config-6s9mt\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.505527 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-var-run-ovn\") pod \"ovn-controller-22dwz-config-6s9mt\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.505651 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94526b19-793f-4ec9-9a80-8a7d4f8cab87-config-data\") pod \"glance-db-sync-6vlxg\" (UID: \"94526b19-793f-4ec9-9a80-8a7d4f8cab87\") " pod="openstack/glance-db-sync-6vlxg" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.505731 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-scripts\") pod \"ovn-controller-22dwz-config-6s9mt\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.505901 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94526b19-793f-4ec9-9a80-8a7d4f8cab87-combined-ca-bundle\") pod \"glance-db-sync-6vlxg\" (UID: \"94526b19-793f-4ec9-9a80-8a7d4f8cab87\") " pod="openstack/glance-db-sync-6vlxg" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.505940 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/94526b19-793f-4ec9-9a80-8a7d4f8cab87-db-sync-config-data\") pod \"glance-db-sync-6vlxg\" (UID: \"94526b19-793f-4ec9-9a80-8a7d4f8cab87\") " pod="openstack/glance-db-sync-6vlxg" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.510504 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94526b19-793f-4ec9-9a80-8a7d4f8cab87-config-data\") pod \"glance-db-sync-6vlxg\" (UID: \"94526b19-793f-4ec9-9a80-8a7d4f8cab87\") " pod="openstack/glance-db-sync-6vlxg" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.510929 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94526b19-793f-4ec9-9a80-8a7d4f8cab87-combined-ca-bundle\") pod \"glance-db-sync-6vlxg\" (UID: \"94526b19-793f-4ec9-9a80-8a7d4f8cab87\") " pod="openstack/glance-db-sync-6vlxg" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.523270 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/94526b19-793f-4ec9-9a80-8a7d4f8cab87-db-sync-config-data\") pod \"glance-db-sync-6vlxg\" (UID: \"94526b19-793f-4ec9-9a80-8a7d4f8cab87\") " pod="openstack/glance-db-sync-6vlxg" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.523878 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt5pp\" (UniqueName: \"kubernetes.io/projected/94526b19-793f-4ec9-9a80-8a7d4f8cab87-kube-api-access-rt5pp\") pod \"glance-db-sync-6vlxg\" (UID: \"94526b19-793f-4ec9-9a80-8a7d4f8cab87\") " pod="openstack/glance-db-sync-6vlxg" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.570439 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6vlxg" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.607250 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-var-run\") pod \"ovn-controller-22dwz-config-6s9mt\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.607350 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-var-log-ovn\") pod \"ovn-controller-22dwz-config-6s9mt\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.607389 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-additional-scripts\") pod \"ovn-controller-22dwz-config-6s9mt\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.607417 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-var-run-ovn\") pod \"ovn-controller-22dwz-config-6s9mt\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.607491 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-scripts\") pod \"ovn-controller-22dwz-config-6s9mt\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.607582 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szv5t\" (UniqueName: \"kubernetes.io/projected/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-kube-api-access-szv5t\") pod \"ovn-controller-22dwz-config-6s9mt\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.607584 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-var-run\") pod \"ovn-controller-22dwz-config-6s9mt\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.608388 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-additional-scripts\") pod \"ovn-controller-22dwz-config-6s9mt\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.608601 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-var-run-ovn\") pod \"ovn-controller-22dwz-config-6s9mt\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.608663 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-var-log-ovn\") pod \"ovn-controller-22dwz-config-6s9mt\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.610034 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-scripts\") pod \"ovn-controller-22dwz-config-6s9mt\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.636775 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szv5t\" (UniqueName: \"kubernetes.io/projected/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-kube-api-access-szv5t\") pod \"ovn-controller-22dwz-config-6s9mt\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.695509 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:32 crc kubenswrapper[4934]: I1227 08:03:32.907214 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-22dwz" Dec 27 08:03:33 crc kubenswrapper[4934]: I1227 08:03:33.352991 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-6vlxg"] Dec 27 08:03:33 crc kubenswrapper[4934]: W1227 08:03:33.357281 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94526b19_793f_4ec9_9a80_8a7d4f8cab87.slice/crio-5ab7f8abc9597cb13455904036db82aad1a788a11cb993a7a718c8afa4b94b2f WatchSource:0}: Error finding container 5ab7f8abc9597cb13455904036db82aad1a788a11cb993a7a718c8afa4b94b2f: Status 404 returned error can't find the container with id 5ab7f8abc9597cb13455904036db82aad1a788a11cb993a7a718c8afa4b94b2f Dec 27 08:03:33 crc kubenswrapper[4934]: W1227 08:03:33.427092 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf86e8ff6_0db6_4708_8efa_6e39c37b5ec8.slice/crio-72dcf8c106586fb318f373e2d2aac9976c1be72274e8a398c082351249060485 WatchSource:0}: Error finding container 72dcf8c106586fb318f373e2d2aac9976c1be72274e8a398c082351249060485: Status 404 returned error can't find the container with id 72dcf8c106586fb318f373e2d2aac9976c1be72274e8a398c082351249060485 Dec 27 08:03:33 crc kubenswrapper[4934]: I1227 08:03:33.429592 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-22dwz-config-6s9mt"] Dec 27 08:03:33 crc kubenswrapper[4934]: I1227 08:03:33.493638 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a7c7644-3523-4533-8e7a-39652454a9a0" path="/var/lib/kubelet/pods/8a7c7644-3523-4533-8e7a-39652454a9a0/volumes" Dec 27 08:03:33 crc kubenswrapper[4934]: I1227 08:03:33.795036 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6vlxg" event={"ID":"94526b19-793f-4ec9-9a80-8a7d4f8cab87","Type":"ContainerStarted","Data":"5ab7f8abc9597cb13455904036db82aad1a788a11cb993a7a718c8afa4b94b2f"} Dec 27 08:03:33 crc kubenswrapper[4934]: I1227 08:03:33.796761 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-22dwz-config-6s9mt" event={"ID":"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8","Type":"ContainerStarted","Data":"f7124ee43fecbe067924a0a5202f619026d658a023e3c1209dc7581a192ef658"} Dec 27 08:03:33 crc kubenswrapper[4934]: I1227 08:03:33.796819 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-22dwz-config-6s9mt" event={"ID":"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8","Type":"ContainerStarted","Data":"72dcf8c106586fb318f373e2d2aac9976c1be72274e8a398c082351249060485"} Dec 27 08:03:33 crc kubenswrapper[4934]: I1227 08:03:33.826766 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-22dwz-config-6s9mt" podStartSLOduration=1.826744473 podStartE2EDuration="1.826744473s" podCreationTimestamp="2025-12-27 08:03:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:03:33.821394358 +0000 UTC m=+1274.641834952" watchObservedRunningTime="2025-12-27 08:03:33.826744473 +0000 UTC m=+1274.647185067" Dec 27 08:03:34 crc kubenswrapper[4934]: I1227 08:03:34.256550 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-bef3-account-create-qm6cg"] Dec 27 08:03:34 crc kubenswrapper[4934]: I1227 08:03:34.258427 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-bef3-account-create-qm6cg" Dec 27 08:03:34 crc kubenswrapper[4934]: I1227 08:03:34.260644 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-cell1-db-secret" Dec 27 08:03:34 crc kubenswrapper[4934]: I1227 08:03:34.270532 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-bef3-account-create-qm6cg"] Dec 27 08:03:34 crc kubenswrapper[4934]: I1227 08:03:34.345401 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6sf2\" (UniqueName: \"kubernetes.io/projected/c9f93658-b339-4655-841b-bebfba9ad231-kube-api-access-s6sf2\") pod \"mysqld-exporter-bef3-account-create-qm6cg\" (UID: \"c9f93658-b339-4655-841b-bebfba9ad231\") " pod="openstack/mysqld-exporter-bef3-account-create-qm6cg" Dec 27 08:03:34 crc kubenswrapper[4934]: I1227 08:03:34.447148 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6sf2\" (UniqueName: \"kubernetes.io/projected/c9f93658-b339-4655-841b-bebfba9ad231-kube-api-access-s6sf2\") pod \"mysqld-exporter-bef3-account-create-qm6cg\" (UID: \"c9f93658-b339-4655-841b-bebfba9ad231\") " pod="openstack/mysqld-exporter-bef3-account-create-qm6cg" Dec 27 08:03:34 crc kubenswrapper[4934]: I1227 08:03:34.466332 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6sf2\" (UniqueName: \"kubernetes.io/projected/c9f93658-b339-4655-841b-bebfba9ad231-kube-api-access-s6sf2\") pod \"mysqld-exporter-bef3-account-create-qm6cg\" (UID: \"c9f93658-b339-4655-841b-bebfba9ad231\") " pod="openstack/mysqld-exporter-bef3-account-create-qm6cg" Dec 27 08:03:34 crc kubenswrapper[4934]: I1227 08:03:34.572611 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-bef3-account-create-qm6cg" Dec 27 08:03:34 crc kubenswrapper[4934]: I1227 08:03:34.816241 4934 generic.go:334] "Generic (PLEG): container finished" podID="f86e8ff6-0db6-4708-8efa-6e39c37b5ec8" containerID="f7124ee43fecbe067924a0a5202f619026d658a023e3c1209dc7581a192ef658" exitCode=0 Dec 27 08:03:34 crc kubenswrapper[4934]: I1227 08:03:34.816332 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-22dwz-config-6s9mt" event={"ID":"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8","Type":"ContainerDied","Data":"f7124ee43fecbe067924a0a5202f619026d658a023e3c1209dc7581a192ef658"} Dec 27 08:03:35 crc kubenswrapper[4934]: I1227 08:03:35.066124 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-bef3-account-create-qm6cg"] Dec 27 08:03:35 crc kubenswrapper[4934]: I1227 08:03:35.830014 4934 generic.go:334] "Generic (PLEG): container finished" podID="c9f93658-b339-4655-841b-bebfba9ad231" containerID="58eec049f82987e1a6707f16ad161c530040c78d71a3c2b424895f1b1549ea3c" exitCode=0 Dec 27 08:03:35 crc kubenswrapper[4934]: I1227 08:03:35.830118 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-bef3-account-create-qm6cg" event={"ID":"c9f93658-b339-4655-841b-bebfba9ad231","Type":"ContainerDied","Data":"58eec049f82987e1a6707f16ad161c530040c78d71a3c2b424895f1b1549ea3c"} Dec 27 08:03:35 crc kubenswrapper[4934]: I1227 08:03:35.830485 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-bef3-account-create-qm6cg" event={"ID":"c9f93658-b339-4655-841b-bebfba9ad231","Type":"ContainerStarted","Data":"4e04199cdcdceb060afd5540465288715ec0f12103ddd06885de67ff4cbf858b"} Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.198832 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.314301 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-var-run-ovn\") pod \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.314375 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-additional-scripts\") pod \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.314446 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-scripts\") pod \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.314516 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szv5t\" (UniqueName: \"kubernetes.io/projected/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-kube-api-access-szv5t\") pod \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.314572 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-var-run\") pod \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.314710 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-var-log-ovn\") pod \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\" (UID: \"f86e8ff6-0db6-4708-8efa-6e39c37b5ec8\") " Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.315121 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "f86e8ff6-0db6-4708-8efa-6e39c37b5ec8" (UID: "f86e8ff6-0db6-4708-8efa-6e39c37b5ec8"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.315150 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "f86e8ff6-0db6-4708-8efa-6e39c37b5ec8" (UID: "f86e8ff6-0db6-4708-8efa-6e39c37b5ec8"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.315173 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-var-run" (OuterVolumeSpecName: "var-run") pod "f86e8ff6-0db6-4708-8efa-6e39c37b5ec8" (UID: "f86e8ff6-0db6-4708-8efa-6e39c37b5ec8"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.315173 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "f86e8ff6-0db6-4708-8efa-6e39c37b5ec8" (UID: "f86e8ff6-0db6-4708-8efa-6e39c37b5ec8"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.316154 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-scripts" (OuterVolumeSpecName: "scripts") pod "f86e8ff6-0db6-4708-8efa-6e39c37b5ec8" (UID: "f86e8ff6-0db6-4708-8efa-6e39c37b5ec8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.335377 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-kube-api-access-szv5t" (OuterVolumeSpecName: "kube-api-access-szv5t") pod "f86e8ff6-0db6-4708-8efa-6e39c37b5ec8" (UID: "f86e8ff6-0db6-4708-8efa-6e39c37b5ec8"). InnerVolumeSpecName "kube-api-access-szv5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.416946 4934 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.416978 4934 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.416987 4934 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.416997 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.417005 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szv5t\" (UniqueName: \"kubernetes.io/projected/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-kube-api-access-szv5t\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.417015 4934 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8-var-run\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.499285 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-22dwz-config-6s9mt"] Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.506903 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-22dwz-config-6s9mt"] Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.825252 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-etc-swift\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") " pod="openstack/swift-storage-0" Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.840973 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e59489eb-aaa1-4188-b593-d3d3c2f1f412-etc-swift\") pod \"swift-storage-0\" (UID: \"e59489eb-aaa1-4188-b593-d3d3c2f1f412\") " pod="openstack/swift-storage-0" Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.865176 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72dcf8c106586fb318f373e2d2aac9976c1be72274e8a398c082351249060485" Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.865316 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-22dwz-config-6s9mt" Dec 27 08:03:36 crc kubenswrapper[4934]: I1227 08:03:36.951645 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 27 08:03:37 crc kubenswrapper[4934]: I1227 08:03:37.300501 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-bef3-account-create-qm6cg" Dec 27 08:03:37 crc kubenswrapper[4934]: I1227 08:03:37.435915 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6sf2\" (UniqueName: \"kubernetes.io/projected/c9f93658-b339-4655-841b-bebfba9ad231-kube-api-access-s6sf2\") pod \"c9f93658-b339-4655-841b-bebfba9ad231\" (UID: \"c9f93658-b339-4655-841b-bebfba9ad231\") " Dec 27 08:03:37 crc kubenswrapper[4934]: I1227 08:03:37.440894 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9f93658-b339-4655-841b-bebfba9ad231-kube-api-access-s6sf2" (OuterVolumeSpecName: "kube-api-access-s6sf2") pod "c9f93658-b339-4655-841b-bebfba9ad231" (UID: "c9f93658-b339-4655-841b-bebfba9ad231"). InnerVolumeSpecName "kube-api-access-s6sf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:37 crc kubenswrapper[4934]: I1227 08:03:37.481138 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f86e8ff6-0db6-4708-8efa-6e39c37b5ec8" path="/var/lib/kubelet/pods/f86e8ff6-0db6-4708-8efa-6e39c37b5ec8/volumes" Dec 27 08:03:37 crc kubenswrapper[4934]: I1227 08:03:37.538254 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6sf2\" (UniqueName: \"kubernetes.io/projected/c9f93658-b339-4655-841b-bebfba9ad231-kube-api-access-s6sf2\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:37 crc kubenswrapper[4934]: I1227 08:03:37.555388 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 27 08:03:37 crc kubenswrapper[4934]: I1227 08:03:37.874656 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e59489eb-aaa1-4188-b593-d3d3c2f1f412","Type":"ContainerStarted","Data":"8dae8f62ce3997c8fff9abde8c01f23a819013e9cba85f47d403333179d1b540"} Dec 27 08:03:37 crc kubenswrapper[4934]: I1227 08:03:37.877322 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-bef3-account-create-qm6cg" event={"ID":"c9f93658-b339-4655-841b-bebfba9ad231","Type":"ContainerDied","Data":"4e04199cdcdceb060afd5540465288715ec0f12103ddd06885de67ff4cbf858b"} Dec 27 08:03:37 crc kubenswrapper[4934]: I1227 08:03:37.877370 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e04199cdcdceb060afd5540465288715ec0f12103ddd06885de67ff4cbf858b" Dec 27 08:03:37 crc kubenswrapper[4934]: I1227 08:03:37.877380 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-bef3-account-create-qm6cg" Dec 27 08:03:38 crc kubenswrapper[4934]: I1227 08:03:38.810262 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 27 08:03:38 crc kubenswrapper[4934]: I1227 08:03:38.837128 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-1" podUID="5788f6c6-2500-4c66-af7a-830cde17caa4" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.130:5671: connect: connection refused" Dec 27 08:03:38 crc kubenswrapper[4934]: I1227 08:03:38.852323 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-2" Dec 27 08:03:39 crc kubenswrapper[4934]: I1227 08:03:39.514233 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Dec 27 08:03:39 crc kubenswrapper[4934]: E1227 08:03:39.514814 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f86e8ff6-0db6-4708-8efa-6e39c37b5ec8" containerName="ovn-config" Dec 27 08:03:39 crc kubenswrapper[4934]: I1227 08:03:39.514831 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f86e8ff6-0db6-4708-8efa-6e39c37b5ec8" containerName="ovn-config" Dec 27 08:03:39 crc kubenswrapper[4934]: E1227 08:03:39.514874 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9f93658-b339-4655-841b-bebfba9ad231" containerName="mariadb-account-create" Dec 27 08:03:39 crc kubenswrapper[4934]: I1227 08:03:39.514884 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9f93658-b339-4655-841b-bebfba9ad231" containerName="mariadb-account-create" Dec 27 08:03:39 crc kubenswrapper[4934]: I1227 08:03:39.515125 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9f93658-b339-4655-841b-bebfba9ad231" containerName="mariadb-account-create" Dec 27 08:03:39 crc kubenswrapper[4934]: I1227 08:03:39.515150 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f86e8ff6-0db6-4708-8efa-6e39c37b5ec8" containerName="ovn-config" Dec 27 08:03:39 crc kubenswrapper[4934]: I1227 08:03:39.516001 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 27 08:03:39 crc kubenswrapper[4934]: I1227 08:03:39.517838 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Dec 27 08:03:39 crc kubenswrapper[4934]: I1227 08:03:39.578652 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 27 08:03:39 crc kubenswrapper[4934]: I1227 08:03:39.605380 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lvg9\" (UniqueName: \"kubernetes.io/projected/040f4268-549c-4c0e-885f-cc9a7dcf4e03-kube-api-access-7lvg9\") pod \"mysqld-exporter-0\" (UID: \"040f4268-549c-4c0e-885f-cc9a7dcf4e03\") " pod="openstack/mysqld-exporter-0" Dec 27 08:03:39 crc kubenswrapper[4934]: I1227 08:03:39.605503 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/040f4268-549c-4c0e-885f-cc9a7dcf4e03-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"040f4268-549c-4c0e-885f-cc9a7dcf4e03\") " pod="openstack/mysqld-exporter-0" Dec 27 08:03:39 crc kubenswrapper[4934]: I1227 08:03:39.605623 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/040f4268-549c-4c0e-885f-cc9a7dcf4e03-config-data\") pod \"mysqld-exporter-0\" (UID: \"040f4268-549c-4c0e-885f-cc9a7dcf4e03\") " pod="openstack/mysqld-exporter-0" Dec 27 08:03:39 crc kubenswrapper[4934]: I1227 08:03:39.707998 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/040f4268-549c-4c0e-885f-cc9a7dcf4e03-config-data\") pod \"mysqld-exporter-0\" (UID: \"040f4268-549c-4c0e-885f-cc9a7dcf4e03\") " pod="openstack/mysqld-exporter-0" Dec 27 08:03:39 crc kubenswrapper[4934]: I1227 08:03:39.708099 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lvg9\" (UniqueName: \"kubernetes.io/projected/040f4268-549c-4c0e-885f-cc9a7dcf4e03-kube-api-access-7lvg9\") pod \"mysqld-exporter-0\" (UID: \"040f4268-549c-4c0e-885f-cc9a7dcf4e03\") " pod="openstack/mysqld-exporter-0" Dec 27 08:03:39 crc kubenswrapper[4934]: I1227 08:03:39.708170 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/040f4268-549c-4c0e-885f-cc9a7dcf4e03-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"040f4268-549c-4c0e-885f-cc9a7dcf4e03\") " pod="openstack/mysqld-exporter-0" Dec 27 08:03:39 crc kubenswrapper[4934]: I1227 08:03:39.717512 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/040f4268-549c-4c0e-885f-cc9a7dcf4e03-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"040f4268-549c-4c0e-885f-cc9a7dcf4e03\") " pod="openstack/mysqld-exporter-0" Dec 27 08:03:39 crc kubenswrapper[4934]: I1227 08:03:39.722177 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/040f4268-549c-4c0e-885f-cc9a7dcf4e03-config-data\") pod \"mysqld-exporter-0\" (UID: \"040f4268-549c-4c0e-885f-cc9a7dcf4e03\") " pod="openstack/mysqld-exporter-0" Dec 27 08:03:39 crc kubenswrapper[4934]: I1227 08:03:39.732647 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lvg9\" (UniqueName: \"kubernetes.io/projected/040f4268-549c-4c0e-885f-cc9a7dcf4e03-kube-api-access-7lvg9\") pod \"mysqld-exporter-0\" (UID: \"040f4268-549c-4c0e-885f-cc9a7dcf4e03\") " pod="openstack/mysqld-exporter-0" Dec 27 08:03:39 crc kubenswrapper[4934]: I1227 08:03:39.837291 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 27 08:03:40 crc kubenswrapper[4934]: I1227 08:03:40.272464 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:40 crc kubenswrapper[4934]: I1227 08:03:40.274800 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:41 crc kubenswrapper[4934]: I1227 08:03:41.038286 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:43 crc kubenswrapper[4934]: I1227 08:03:43.508257 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 27 08:03:43 crc kubenswrapper[4934]: I1227 08:03:43.508589 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="92d66221-653f-4837-a733-f02073b6ed70" containerName="prometheus" containerID="cri-o://12b822224b3e8dc94c80b53120e42ed082e0f0e3b3001ded509f8d22b739a9e4" gracePeriod=600 Dec 27 08:03:43 crc kubenswrapper[4934]: I1227 08:03:43.508667 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="92d66221-653f-4837-a733-f02073b6ed70" containerName="config-reloader" containerID="cri-o://46acee5837df489cb9e18ac9daa2f27e318770779025b31476deffa809d0a809" gracePeriod=600 Dec 27 08:03:43 crc kubenswrapper[4934]: I1227 08:03:43.508697 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="92d66221-653f-4837-a733-f02073b6ed70" containerName="thanos-sidecar" containerID="cri-o://80a5156434ae03cae6f2c876737e9a1f41e83e3a35630adcd0abe3e0277e0270" gracePeriod=600 Dec 27 08:03:44 crc kubenswrapper[4934]: I1227 08:03:44.070538 4934 generic.go:334] "Generic (PLEG): container finished" podID="92d66221-653f-4837-a733-f02073b6ed70" containerID="80a5156434ae03cae6f2c876737e9a1f41e83e3a35630adcd0abe3e0277e0270" exitCode=0 Dec 27 08:03:44 crc kubenswrapper[4934]: I1227 08:03:44.070808 4934 generic.go:334] "Generic (PLEG): container finished" podID="92d66221-653f-4837-a733-f02073b6ed70" containerID="46acee5837df489cb9e18ac9daa2f27e318770779025b31476deffa809d0a809" exitCode=0 Dec 27 08:03:44 crc kubenswrapper[4934]: I1227 08:03:44.070817 4934 generic.go:334] "Generic (PLEG): container finished" podID="92d66221-653f-4837-a733-f02073b6ed70" containerID="12b822224b3e8dc94c80b53120e42ed082e0f0e3b3001ded509f8d22b739a9e4" exitCode=0 Dec 27 08:03:44 crc kubenswrapper[4934]: I1227 08:03:44.070680 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92d66221-653f-4837-a733-f02073b6ed70","Type":"ContainerDied","Data":"80a5156434ae03cae6f2c876737e9a1f41e83e3a35630adcd0abe3e0277e0270"} Dec 27 08:03:44 crc kubenswrapper[4934]: I1227 08:03:44.070854 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92d66221-653f-4837-a733-f02073b6ed70","Type":"ContainerDied","Data":"46acee5837df489cb9e18ac9daa2f27e318770779025b31476deffa809d0a809"} Dec 27 08:03:44 crc kubenswrapper[4934]: I1227 08:03:44.070868 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92d66221-653f-4837-a733-f02073b6ed70","Type":"ContainerDied","Data":"12b822224b3e8dc94c80b53120e42ed082e0f0e3b3001ded509f8d22b739a9e4"} Dec 27 08:03:45 crc kubenswrapper[4934]: I1227 08:03:45.273343 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="92d66221-653f-4837-a733-f02073b6ed70" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.139:9090/-/ready\": dial tcp 10.217.0.139:9090: connect: connection refused" Dec 27 08:03:48 crc kubenswrapper[4934]: I1227 08:03:48.835269 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-1" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.289906 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-vcjnz"] Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.291515 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vcjnz" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.305226 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vcjnz"] Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.354690 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w2xl\" (UniqueName: \"kubernetes.io/projected/a4aa74e5-8fd8-45e5-907d-e95f8bf11d37-kube-api-access-5w2xl\") pod \"cinder-db-create-vcjnz\" (UID: \"a4aa74e5-8fd8-45e5-907d-e95f8bf11d37\") " pod="openstack/cinder-db-create-vcjnz" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.379648 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-858rb"] Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.380969 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-858rb" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.404802 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-858rb"] Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.456468 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfw94\" (UniqueName: \"kubernetes.io/projected/47797090-7e7d-40ba-bf3f-a43d18af9283-kube-api-access-bfw94\") pod \"barbican-db-create-858rb\" (UID: \"47797090-7e7d-40ba-bf3f-a43d18af9283\") " pod="openstack/barbican-db-create-858rb" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.456702 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w2xl\" (UniqueName: \"kubernetes.io/projected/a4aa74e5-8fd8-45e5-907d-e95f8bf11d37-kube-api-access-5w2xl\") pod \"cinder-db-create-vcjnz\" (UID: \"a4aa74e5-8fd8-45e5-907d-e95f8bf11d37\") " pod="openstack/cinder-db-create-vcjnz" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.464565 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-dhmx4"] Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.466424 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-dhmx4" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.489025 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w2xl\" (UniqueName: \"kubernetes.io/projected/a4aa74e5-8fd8-45e5-907d-e95f8bf11d37-kube-api-access-5w2xl\") pod \"cinder-db-create-vcjnz\" (UID: \"a4aa74e5-8fd8-45e5-907d-e95f8bf11d37\") " pod="openstack/cinder-db-create-vcjnz" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.498758 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-dhmx4"] Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.558299 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cjxr\" (UniqueName: \"kubernetes.io/projected/06fd3b33-8755-4e16-93df-6ca70836a8f7-kube-api-access-2cjxr\") pod \"heat-db-create-dhmx4\" (UID: \"06fd3b33-8755-4e16-93df-6ca70836a8f7\") " pod="openstack/heat-db-create-dhmx4" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.559600 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfw94\" (UniqueName: \"kubernetes.io/projected/47797090-7e7d-40ba-bf3f-a43d18af9283-kube-api-access-bfw94\") pod \"barbican-db-create-858rb\" (UID: \"47797090-7e7d-40ba-bf3f-a43d18af9283\") " pod="openstack/barbican-db-create-858rb" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.560021 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-cgptv"] Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.561818 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-cgptv" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.566400 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.566611 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-8r767" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.566737 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.567182 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.575895 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-cgptv"] Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.582695 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfw94\" (UniqueName: \"kubernetes.io/projected/47797090-7e7d-40ba-bf3f-a43d18af9283-kube-api-access-bfw94\") pod \"barbican-db-create-858rb\" (UID: \"47797090-7e7d-40ba-bf3f-a43d18af9283\") " pod="openstack/barbican-db-create-858rb" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.609658 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vcjnz" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.661496 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c811bf9-1bb4-4f9b-8967-9a67bb94c520-config-data\") pod \"keystone-db-sync-cgptv\" (UID: \"5c811bf9-1bb4-4f9b-8967-9a67bb94c520\") " pod="openstack/keystone-db-sync-cgptv" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.661613 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cjxr\" (UniqueName: \"kubernetes.io/projected/06fd3b33-8755-4e16-93df-6ca70836a8f7-kube-api-access-2cjxr\") pod \"heat-db-create-dhmx4\" (UID: \"06fd3b33-8755-4e16-93df-6ca70836a8f7\") " pod="openstack/heat-db-create-dhmx4" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.661731 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qs64\" (UniqueName: \"kubernetes.io/projected/5c811bf9-1bb4-4f9b-8967-9a67bb94c520-kube-api-access-5qs64\") pod \"keystone-db-sync-cgptv\" (UID: \"5c811bf9-1bb4-4f9b-8967-9a67bb94c520\") " pod="openstack/keystone-db-sync-cgptv" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.661769 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c811bf9-1bb4-4f9b-8967-9a67bb94c520-combined-ca-bundle\") pod \"keystone-db-sync-cgptv\" (UID: \"5c811bf9-1bb4-4f9b-8967-9a67bb94c520\") " pod="openstack/keystone-db-sync-cgptv" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.667741 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-92xmf"] Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.669154 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-92xmf" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.685951 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-92xmf"] Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.693217 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cjxr\" (UniqueName: \"kubernetes.io/projected/06fd3b33-8755-4e16-93df-6ca70836a8f7-kube-api-access-2cjxr\") pod \"heat-db-create-dhmx4\" (UID: \"06fd3b33-8755-4e16-93df-6ca70836a8f7\") " pod="openstack/heat-db-create-dhmx4" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.702561 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-858rb" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.763602 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8wsw\" (UniqueName: \"kubernetes.io/projected/d745a073-8cd4-463c-ae78-53185e910777-kube-api-access-c8wsw\") pod \"neutron-db-create-92xmf\" (UID: \"d745a073-8cd4-463c-ae78-53185e910777\") " pod="openstack/neutron-db-create-92xmf" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.763653 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qs64\" (UniqueName: \"kubernetes.io/projected/5c811bf9-1bb4-4f9b-8967-9a67bb94c520-kube-api-access-5qs64\") pod \"keystone-db-sync-cgptv\" (UID: \"5c811bf9-1bb4-4f9b-8967-9a67bb94c520\") " pod="openstack/keystone-db-sync-cgptv" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.763683 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c811bf9-1bb4-4f9b-8967-9a67bb94c520-combined-ca-bundle\") pod \"keystone-db-sync-cgptv\" (UID: \"5c811bf9-1bb4-4f9b-8967-9a67bb94c520\") " pod="openstack/keystone-db-sync-cgptv" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.764357 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c811bf9-1bb4-4f9b-8967-9a67bb94c520-config-data\") pod \"keystone-db-sync-cgptv\" (UID: \"5c811bf9-1bb4-4f9b-8967-9a67bb94c520\") " pod="openstack/keystone-db-sync-cgptv" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.767687 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c811bf9-1bb4-4f9b-8967-9a67bb94c520-config-data\") pod \"keystone-db-sync-cgptv\" (UID: \"5c811bf9-1bb4-4f9b-8967-9a67bb94c520\") " pod="openstack/keystone-db-sync-cgptv" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.768574 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c811bf9-1bb4-4f9b-8967-9a67bb94c520-combined-ca-bundle\") pod \"keystone-db-sync-cgptv\" (UID: \"5c811bf9-1bb4-4f9b-8967-9a67bb94c520\") " pod="openstack/keystone-db-sync-cgptv" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.780612 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qs64\" (UniqueName: \"kubernetes.io/projected/5c811bf9-1bb4-4f9b-8967-9a67bb94c520-kube-api-access-5qs64\") pod \"keystone-db-sync-cgptv\" (UID: \"5c811bf9-1bb4-4f9b-8967-9a67bb94c520\") " pod="openstack/keystone-db-sync-cgptv" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.830462 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-dhmx4" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.866402 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8wsw\" (UniqueName: \"kubernetes.io/projected/d745a073-8cd4-463c-ae78-53185e910777-kube-api-access-c8wsw\") pod \"neutron-db-create-92xmf\" (UID: \"d745a073-8cd4-463c-ae78-53185e910777\") " pod="openstack/neutron-db-create-92xmf" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.880701 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8wsw\" (UniqueName: \"kubernetes.io/projected/d745a073-8cd4-463c-ae78-53185e910777-kube-api-access-c8wsw\") pod \"neutron-db-create-92xmf\" (UID: \"d745a073-8cd4-463c-ae78-53185e910777\") " pod="openstack/neutron-db-create-92xmf" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.883747 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-cgptv" Dec 27 08:03:49 crc kubenswrapper[4934]: I1227 08:03:49.985074 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-92xmf" Dec 27 08:03:50 crc kubenswrapper[4934]: I1227 08:03:50.274263 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="92d66221-653f-4837-a733-f02073b6ed70" containerName="prometheus" probeResult="failure" output="Get \"http://10.217.0.139:9090/-/ready\": dial tcp 10.217.0.139:9090: connect: connection refused" Dec 27 08:03:53 crc kubenswrapper[4934]: E1227 08:03:53.264762 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 27 08:03:53 crc kubenswrapper[4934]: E1227 08:03:53.265281 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rt5pp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-6vlxg_openstack(94526b19-793f-4ec9-9a80-8a7d4f8cab87): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:03:53 crc kubenswrapper[4934]: E1227 08:03:53.266931 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-6vlxg" podUID="94526b19-793f-4ec9-9a80-8a7d4f8cab87" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.655915 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.746000 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92d66221-653f-4837-a733-f02073b6ed70-prometheus-metric-storage-rulefiles-0\") pod \"92d66221-653f-4837-a733-f02073b6ed70\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.746099 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92d66221-653f-4837-a733-f02073b6ed70-web-config\") pod \"92d66221-653f-4837-a733-f02073b6ed70\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.746175 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92d66221-653f-4837-a733-f02073b6ed70-tls-assets\") pod \"92d66221-653f-4837-a733-f02073b6ed70\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.746199 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/92d66221-653f-4837-a733-f02073b6ed70-prometheus-metric-storage-rulefiles-2\") pod \"92d66221-653f-4837-a733-f02073b6ed70\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.746228 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/92d66221-653f-4837-a733-f02073b6ed70-prometheus-metric-storage-rulefiles-1\") pod \"92d66221-653f-4837-a733-f02073b6ed70\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.746413 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b\") pod \"92d66221-653f-4837-a733-f02073b6ed70\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.746453 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92d66221-653f-4837-a733-f02073b6ed70-thanos-prometheus-http-client-file\") pod \"92d66221-653f-4837-a733-f02073b6ed70\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.746482 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92d66221-653f-4837-a733-f02073b6ed70-config-out\") pod \"92d66221-653f-4837-a733-f02073b6ed70\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.746506 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vr285\" (UniqueName: \"kubernetes.io/projected/92d66221-653f-4837-a733-f02073b6ed70-kube-api-access-vr285\") pod \"92d66221-653f-4837-a733-f02073b6ed70\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.746596 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/92d66221-653f-4837-a733-f02073b6ed70-config\") pod \"92d66221-653f-4837-a733-f02073b6ed70\" (UID: \"92d66221-653f-4837-a733-f02073b6ed70\") " Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.748188 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92d66221-653f-4837-a733-f02073b6ed70-prometheus-metric-storage-rulefiles-1" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-1") pod "92d66221-653f-4837-a733-f02073b6ed70" (UID: "92d66221-653f-4837-a733-f02073b6ed70"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.750689 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92d66221-653f-4837-a733-f02073b6ed70-prometheus-metric-storage-rulefiles-2" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-2") pod "92d66221-653f-4837-a733-f02073b6ed70" (UID: "92d66221-653f-4837-a733-f02073b6ed70"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-2". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.751271 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92d66221-653f-4837-a733-f02073b6ed70-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "92d66221-653f-4837-a733-f02073b6ed70" (UID: "92d66221-653f-4837-a733-f02073b6ed70"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.761456 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92d66221-653f-4837-a733-f02073b6ed70-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "92d66221-653f-4837-a733-f02073b6ed70" (UID: "92d66221-653f-4837-a733-f02073b6ed70"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.761607 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92d66221-653f-4837-a733-f02073b6ed70-kube-api-access-vr285" (OuterVolumeSpecName: "kube-api-access-vr285") pod "92d66221-653f-4837-a733-f02073b6ed70" (UID: "92d66221-653f-4837-a733-f02073b6ed70"). InnerVolumeSpecName "kube-api-access-vr285". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.762171 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92d66221-653f-4837-a733-f02073b6ed70-config-out" (OuterVolumeSpecName: "config-out") pod "92d66221-653f-4837-a733-f02073b6ed70" (UID: "92d66221-653f-4837-a733-f02073b6ed70"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.762674 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d66221-653f-4837-a733-f02073b6ed70-config" (OuterVolumeSpecName: "config") pod "92d66221-653f-4837-a733-f02073b6ed70" (UID: "92d66221-653f-4837-a733-f02073b6ed70"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.762713 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d66221-653f-4837-a733-f02073b6ed70-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "92d66221-653f-4837-a733-f02073b6ed70" (UID: "92d66221-653f-4837-a733-f02073b6ed70"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.838716 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d66221-653f-4837-a733-f02073b6ed70-web-config" (OuterVolumeSpecName: "web-config") pod "92d66221-653f-4837-a733-f02073b6ed70" (UID: "92d66221-653f-4837-a733-f02073b6ed70"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.839469 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "92d66221-653f-4837-a733-f02073b6ed70" (UID: "92d66221-653f-4837-a733-f02073b6ed70"). InnerVolumeSpecName "pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.849210 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-858rb"] Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.850992 4934 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/92d66221-653f-4837-a733-f02073b6ed70-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.851138 4934 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/92d66221-653f-4837-a733-f02073b6ed70-prometheus-metric-storage-rulefiles-2\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.851250 4934 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/92d66221-653f-4837-a733-f02073b6ed70-prometheus-metric-storage-rulefiles-1\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.851349 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b\") on node \"crc\" " Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.851440 4934 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/92d66221-653f-4837-a733-f02073b6ed70-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.851526 4934 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/92d66221-653f-4837-a733-f02073b6ed70-config-out\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.851633 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vr285\" (UniqueName: \"kubernetes.io/projected/92d66221-653f-4837-a733-f02073b6ed70-kube-api-access-vr285\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.851714 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/92d66221-653f-4837-a733-f02073b6ed70-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.851791 4934 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/92d66221-653f-4837-a733-f02073b6ed70-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.851865 4934 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/92d66221-653f-4837-a733-f02073b6ed70-web-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.932986 4934 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.933147 4934 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b") on node "crc" Dec 27 08:03:53 crc kubenswrapper[4934]: I1227 08:03:53.956384 4934 reconciler_common.go:293] "Volume detached for volume \"pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.097090 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-dhmx4"] Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.194806 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"92d66221-653f-4837-a733-f02073b6ed70","Type":"ContainerDied","Data":"7e838cf66fc939ea1244896217720d1205d8a0cdeb45401e1a154aabbe99ae2c"} Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.194896 4934 scope.go:117] "RemoveContainer" containerID="80a5156434ae03cae6f2c876737e9a1f41e83e3a35630adcd0abe3e0277e0270" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.195253 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.199973 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-dhmx4" event={"ID":"06fd3b33-8755-4e16-93df-6ca70836a8f7","Type":"ContainerStarted","Data":"f4fe7ca0e7a49870e4d0c6df645702587f7c1f716e85e0fa093c20214db9192c"} Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.202993 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-858rb" event={"ID":"47797090-7e7d-40ba-bf3f-a43d18af9283","Type":"ContainerStarted","Data":"9c7a962c5fbfa08d7139f70735d236be10782026de00fe694b1949d1d850814b"} Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.203039 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-858rb" event={"ID":"47797090-7e7d-40ba-bf3f-a43d18af9283","Type":"ContainerStarted","Data":"bd48b38eda25f86f8a7bfcc5bf6f79ad1834faeaf57e2443606ff96d60e303ed"} Dec 27 08:03:54 crc kubenswrapper[4934]: E1227 08:03:54.206047 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-6vlxg" podUID="94526b19-793f-4ec9-9a80-8a7d4f8cab87" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.302479 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.309736 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-858rb" podStartSLOduration=5.309716792 podStartE2EDuration="5.309716792s" podCreationTimestamp="2025-12-27 08:03:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:03:54.251657776 +0000 UTC m=+1295.072098380" watchObservedRunningTime="2025-12-27 08:03:54.309716792 +0000 UTC m=+1295.130157386" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.320730 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-cgptv"] Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.331962 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-92xmf"] Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.342371 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vcjnz"] Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.363037 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.382218 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.392500 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 27 08:03:54 crc kubenswrapper[4934]: E1227 08:03:54.392991 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d66221-653f-4837-a733-f02073b6ed70" containerName="init-config-reloader" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.393010 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d66221-653f-4837-a733-f02073b6ed70" containerName="init-config-reloader" Dec 27 08:03:54 crc kubenswrapper[4934]: E1227 08:03:54.393018 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d66221-653f-4837-a733-f02073b6ed70" containerName="prometheus" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.393023 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d66221-653f-4837-a733-f02073b6ed70" containerName="prometheus" Dec 27 08:03:54 crc kubenswrapper[4934]: E1227 08:03:54.393035 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d66221-653f-4837-a733-f02073b6ed70" containerName="thanos-sidecar" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.393041 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d66221-653f-4837-a733-f02073b6ed70" containerName="thanos-sidecar" Dec 27 08:03:54 crc kubenswrapper[4934]: E1227 08:03:54.393064 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d66221-653f-4837-a733-f02073b6ed70" containerName="config-reloader" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.393070 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d66221-653f-4837-a733-f02073b6ed70" containerName="config-reloader" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.393287 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d66221-653f-4837-a733-f02073b6ed70" containerName="prometheus" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.393303 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d66221-653f-4837-a733-f02073b6ed70" containerName="config-reloader" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.393321 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d66221-653f-4837-a733-f02073b6ed70" containerName="thanos-sidecar" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.395546 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.397479 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.399791 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-1" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.399862 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.400159 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-2" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.400400 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.400587 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-9q2gk" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.400703 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.400791 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.406698 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.412636 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.471677 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.471735 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.471773 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.471814 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-config\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.471835 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.471857 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wdxt\" (UniqueName: \"kubernetes.io/projected/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-kube-api-access-5wdxt\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.471897 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.471916 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.471932 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.471963 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.471993 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.472025 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.472048 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.573469 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.573720 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.573766 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.573806 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.573845 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.573887 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-config\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.573906 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.573930 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wdxt\" (UniqueName: \"kubernetes.io/projected/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-kube-api-access-5wdxt\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.573968 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.573988 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.574005 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.574035 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.574065 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.575181 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.575181 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.576908 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.578094 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.578130 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b43fafdc1175ed3c7018bef2bd4eb270b5f0d66d6e72f88cdbf5483bdbd6cc4f/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.578797 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-config\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.582428 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.582713 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.583459 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.583549 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.584214 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.584400 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.587327 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.590789 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wdxt\" (UniqueName: \"kubernetes.io/projected/e26e3510-3438-4cb9-8d00-f2d0fdf33ef5-kube-api-access-5wdxt\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.622106 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd2f587b-10d7-437b-9200-4c9b8b70b57b\") pod \"prometheus-metric-storage-0\" (UID: \"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5\") " pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.742048 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.754714 4934 scope.go:117] "RemoveContainer" containerID="46acee5837df489cb9e18ac9daa2f27e318770779025b31476deffa809d0a809" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.837909 4934 scope.go:117] "RemoveContainer" containerID="12b822224b3e8dc94c80b53120e42ed082e0f0e3b3001ded509f8d22b739a9e4" Dec 27 08:03:54 crc kubenswrapper[4934]: I1227 08:03:54.861281 4934 scope.go:117] "RemoveContainer" containerID="6984265a9dcba06e22bdc66f68ad44c0acc9f56b5319d26f0929c8d549b57ab3" Dec 27 08:03:55 crc kubenswrapper[4934]: I1227 08:03:55.250190 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e59489eb-aaa1-4188-b593-d3d3c2f1f412","Type":"ContainerStarted","Data":"3ae7c6da04fb541b150fa52656faf9a13a2ac23237fd074f2b6aec959e479190"} Dec 27 08:03:55 crc kubenswrapper[4934]: I1227 08:03:55.253530 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-cgptv" event={"ID":"5c811bf9-1bb4-4f9b-8967-9a67bb94c520","Type":"ContainerStarted","Data":"67260a88d2dca68520e8233eac025f97a21ce3ac8129d3e841f4f7a3b4bce8d5"} Dec 27 08:03:55 crc kubenswrapper[4934]: I1227 08:03:55.261825 4934 generic.go:334] "Generic (PLEG): container finished" podID="47797090-7e7d-40ba-bf3f-a43d18af9283" containerID="9c7a962c5fbfa08d7139f70735d236be10782026de00fe694b1949d1d850814b" exitCode=0 Dec 27 08:03:55 crc kubenswrapper[4934]: I1227 08:03:55.262124 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-858rb" event={"ID":"47797090-7e7d-40ba-bf3f-a43d18af9283","Type":"ContainerDied","Data":"9c7a962c5fbfa08d7139f70735d236be10782026de00fe694b1949d1d850814b"} Dec 27 08:03:55 crc kubenswrapper[4934]: I1227 08:03:55.267857 4934 generic.go:334] "Generic (PLEG): container finished" podID="d745a073-8cd4-463c-ae78-53185e910777" containerID="6f9daf1e58f583c4f515329a771747e6c093a7c929ef3cdd821143d93d088151" exitCode=0 Dec 27 08:03:55 crc kubenswrapper[4934]: I1227 08:03:55.267952 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-92xmf" event={"ID":"d745a073-8cd4-463c-ae78-53185e910777","Type":"ContainerDied","Data":"6f9daf1e58f583c4f515329a771747e6c093a7c929ef3cdd821143d93d088151"} Dec 27 08:03:55 crc kubenswrapper[4934]: I1227 08:03:55.268007 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-92xmf" event={"ID":"d745a073-8cd4-463c-ae78-53185e910777","Type":"ContainerStarted","Data":"f8042cb49b00fbacf2ade54fa04e488241da5b9afe0363ad83be8ec53c98a457"} Dec 27 08:03:55 crc kubenswrapper[4934]: I1227 08:03:55.269770 4934 generic.go:334] "Generic (PLEG): container finished" podID="a4aa74e5-8fd8-45e5-907d-e95f8bf11d37" containerID="39b700197bd5740bb3755b1a69f4fa0e9e21e689c54f2288a3f7d636931af969" exitCode=0 Dec 27 08:03:55 crc kubenswrapper[4934]: I1227 08:03:55.269805 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vcjnz" event={"ID":"a4aa74e5-8fd8-45e5-907d-e95f8bf11d37","Type":"ContainerDied","Data":"39b700197bd5740bb3755b1a69f4fa0e9e21e689c54f2288a3f7d636931af969"} Dec 27 08:03:55 crc kubenswrapper[4934]: I1227 08:03:55.269843 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vcjnz" event={"ID":"a4aa74e5-8fd8-45e5-907d-e95f8bf11d37","Type":"ContainerStarted","Data":"d2abe55ddc8d6f84d792854eb40957a4902dc2afe103310eba96fcce8f6593f2"} Dec 27 08:03:55 crc kubenswrapper[4934]: I1227 08:03:55.272377 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 27 08:03:55 crc kubenswrapper[4934]: I1227 08:03:55.285004 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"040f4268-549c-4c0e-885f-cc9a7dcf4e03","Type":"ContainerStarted","Data":"5d736ec8cff677b28af8405f06f025faf4d1c49e10784b46224ac48aa00d8a5b"} Dec 27 08:03:55 crc kubenswrapper[4934]: I1227 08:03:55.288888 4934 generic.go:334] "Generic (PLEG): container finished" podID="06fd3b33-8755-4e16-93df-6ca70836a8f7" containerID="afb8cdf9e621a319d8617a7c959615983ed7dacecc069a07889905c69202ed0d" exitCode=0 Dec 27 08:03:55 crc kubenswrapper[4934]: I1227 08:03:55.288966 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-dhmx4" event={"ID":"06fd3b33-8755-4e16-93df-6ca70836a8f7","Type":"ContainerDied","Data":"afb8cdf9e621a319d8617a7c959615983ed7dacecc069a07889905c69202ed0d"} Dec 27 08:03:55 crc kubenswrapper[4934]: I1227 08:03:55.481848 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92d66221-653f-4837-a733-f02073b6ed70" path="/var/lib/kubelet/pods/92d66221-653f-4837-a733-f02073b6ed70/volumes" Dec 27 08:03:56 crc kubenswrapper[4934]: I1227 08:03:56.309376 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e59489eb-aaa1-4188-b593-d3d3c2f1f412","Type":"ContainerStarted","Data":"b4c6d45d6cbea7b870c683233a23f11142ebd8b15bfd976ec6f5b255d1709e40"} Dec 27 08:03:56 crc kubenswrapper[4934]: I1227 08:03:56.309669 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e59489eb-aaa1-4188-b593-d3d3c2f1f412","Type":"ContainerStarted","Data":"8b48ce840b3545a1c98b9fc9d2f7b0bec9cad4bf5268e001708e4078b6dc77f0"} Dec 27 08:03:56 crc kubenswrapper[4934]: I1227 08:03:56.310818 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5","Type":"ContainerStarted","Data":"23d730103c4c8accdcee5abed3d178e505d7fb365ad3761cbaf04abb484ed4ac"} Dec 27 08:03:56 crc kubenswrapper[4934]: I1227 08:03:56.985302 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-858rb" Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.083350 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfw94\" (UniqueName: \"kubernetes.io/projected/47797090-7e7d-40ba-bf3f-a43d18af9283-kube-api-access-bfw94\") pod \"47797090-7e7d-40ba-bf3f-a43d18af9283\" (UID: \"47797090-7e7d-40ba-bf3f-a43d18af9283\") " Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.093639 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vcjnz" Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.093932 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47797090-7e7d-40ba-bf3f-a43d18af9283-kube-api-access-bfw94" (OuterVolumeSpecName: "kube-api-access-bfw94") pod "47797090-7e7d-40ba-bf3f-a43d18af9283" (UID: "47797090-7e7d-40ba-bf3f-a43d18af9283"). InnerVolumeSpecName "kube-api-access-bfw94". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.144652 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-dhmx4" Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.185482 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfw94\" (UniqueName: \"kubernetes.io/projected/47797090-7e7d-40ba-bf3f-a43d18af9283-kube-api-access-bfw94\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.211887 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-92xmf" Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.287289 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5w2xl\" (UniqueName: \"kubernetes.io/projected/a4aa74e5-8fd8-45e5-907d-e95f8bf11d37-kube-api-access-5w2xl\") pod \"a4aa74e5-8fd8-45e5-907d-e95f8bf11d37\" (UID: \"a4aa74e5-8fd8-45e5-907d-e95f8bf11d37\") " Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.287417 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cjxr\" (UniqueName: \"kubernetes.io/projected/06fd3b33-8755-4e16-93df-6ca70836a8f7-kube-api-access-2cjxr\") pod \"06fd3b33-8755-4e16-93df-6ca70836a8f7\" (UID: \"06fd3b33-8755-4e16-93df-6ca70836a8f7\") " Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.323188 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-858rb" event={"ID":"47797090-7e7d-40ba-bf3f-a43d18af9283","Type":"ContainerDied","Data":"bd48b38eda25f86f8a7bfcc5bf6f79ad1834faeaf57e2443606ff96d60e303ed"} Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.323214 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-858rb" Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.323228 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd48b38eda25f86f8a7bfcc5bf6f79ad1834faeaf57e2443606ff96d60e303ed" Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.325262 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-92xmf" Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.325267 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-92xmf" event={"ID":"d745a073-8cd4-463c-ae78-53185e910777","Type":"ContainerDied","Data":"f8042cb49b00fbacf2ade54fa04e488241da5b9afe0363ad83be8ec53c98a457"} Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.325340 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8042cb49b00fbacf2ade54fa04e488241da5b9afe0363ad83be8ec53c98a457" Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.327536 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vcjnz" Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.327568 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vcjnz" event={"ID":"a4aa74e5-8fd8-45e5-907d-e95f8bf11d37","Type":"ContainerDied","Data":"d2abe55ddc8d6f84d792854eb40957a4902dc2afe103310eba96fcce8f6593f2"} Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.327603 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2abe55ddc8d6f84d792854eb40957a4902dc2afe103310eba96fcce8f6593f2" Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.329199 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"040f4268-549c-4c0e-885f-cc9a7dcf4e03","Type":"ContainerStarted","Data":"0c3314af0879f460f0bbef2e66c9195a567ab4265adce822d865320da4505fa7"} Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.333254 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-dhmx4" event={"ID":"06fd3b33-8755-4e16-93df-6ca70836a8f7","Type":"ContainerDied","Data":"f4fe7ca0e7a49870e4d0c6df645702587f7c1f716e85e0fa093c20214db9192c"} Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.333299 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4fe7ca0e7a49870e4d0c6df645702587f7c1f716e85e0fa093c20214db9192c" Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.333330 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-dhmx4" Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.336057 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e59489eb-aaa1-4188-b593-d3d3c2f1f412","Type":"ContainerStarted","Data":"7e4ca2656fd014047c680c307a34f9e960fb8516aada381ba82a897dd9c0aa59"} Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.339298 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4aa74e5-8fd8-45e5-907d-e95f8bf11d37-kube-api-access-5w2xl" (OuterVolumeSpecName: "kube-api-access-5w2xl") pod "a4aa74e5-8fd8-45e5-907d-e95f8bf11d37" (UID: "a4aa74e5-8fd8-45e5-907d-e95f8bf11d37"). InnerVolumeSpecName "kube-api-access-5w2xl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.339373 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06fd3b33-8755-4e16-93df-6ca70836a8f7-kube-api-access-2cjxr" (OuterVolumeSpecName: "kube-api-access-2cjxr") pod "06fd3b33-8755-4e16-93df-6ca70836a8f7" (UID: "06fd3b33-8755-4e16-93df-6ca70836a8f7"). InnerVolumeSpecName "kube-api-access-2cjxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.356368 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=16.768714938 podStartE2EDuration="18.356313s" podCreationTimestamp="2025-12-27 08:03:39 +0000 UTC" firstStartedPulling="2025-12-27 08:03:54.756070918 +0000 UTC m=+1295.576511512" lastFinishedPulling="2025-12-27 08:03:56.34366898 +0000 UTC m=+1297.164109574" observedRunningTime="2025-12-27 08:03:57.349425598 +0000 UTC m=+1298.169866202" watchObservedRunningTime="2025-12-27 08:03:57.356313 +0000 UTC m=+1298.176753604" Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.389431 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8wsw\" (UniqueName: \"kubernetes.io/projected/d745a073-8cd4-463c-ae78-53185e910777-kube-api-access-c8wsw\") pod \"d745a073-8cd4-463c-ae78-53185e910777\" (UID: \"d745a073-8cd4-463c-ae78-53185e910777\") " Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.390823 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5w2xl\" (UniqueName: \"kubernetes.io/projected/a4aa74e5-8fd8-45e5-907d-e95f8bf11d37-kube-api-access-5w2xl\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.390864 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cjxr\" (UniqueName: \"kubernetes.io/projected/06fd3b33-8755-4e16-93df-6ca70836a8f7-kube-api-access-2cjxr\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.392748 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d745a073-8cd4-463c-ae78-53185e910777-kube-api-access-c8wsw" (OuterVolumeSpecName: "kube-api-access-c8wsw") pod "d745a073-8cd4-463c-ae78-53185e910777" (UID: "d745a073-8cd4-463c-ae78-53185e910777"). InnerVolumeSpecName "kube-api-access-c8wsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:03:57 crc kubenswrapper[4934]: I1227 08:03:57.492635 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8wsw\" (UniqueName: \"kubernetes.io/projected/d745a073-8cd4-463c-ae78-53185e910777-kube-api-access-c8wsw\") on node \"crc\" DevicePath \"\"" Dec 27 08:03:59 crc kubenswrapper[4934]: I1227 08:03:59.364562 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5","Type":"ContainerStarted","Data":"a6882bd1ae74e1d249c2ec243a06cfe7ba98da11d0e0483844d568496945b12a"} Dec 27 08:04:01 crc kubenswrapper[4934]: I1227 08:04:01.403767 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e59489eb-aaa1-4188-b593-d3d3c2f1f412","Type":"ContainerStarted","Data":"66873477ffd3efd5c0464127b426e1ae07116fde6b636c9e371f943dc46cad8e"} Dec 27 08:04:01 crc kubenswrapper[4934]: I1227 08:04:01.404408 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e59489eb-aaa1-4188-b593-d3d3c2f1f412","Type":"ContainerStarted","Data":"2efdbeaedd360794677e8984fc38d9dd953a323bc7391363fbc399a2342d14ba"} Dec 27 08:04:01 crc kubenswrapper[4934]: I1227 08:04:01.419454 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-cgptv" event={"ID":"5c811bf9-1bb4-4f9b-8967-9a67bb94c520","Type":"ContainerStarted","Data":"cf486a099c077f6365252affac7ca7a4613758c7406362dc41872cda5fa53012"} Dec 27 08:04:01 crc kubenswrapper[4934]: I1227 08:04:01.450068 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-cgptv" podStartSLOduration=6.459193846 podStartE2EDuration="12.450050195s" podCreationTimestamp="2025-12-27 08:03:49 +0000 UTC" firstStartedPulling="2025-12-27 08:03:54.694500464 +0000 UTC m=+1295.514941058" lastFinishedPulling="2025-12-27 08:04:00.685356813 +0000 UTC m=+1301.505797407" observedRunningTime="2025-12-27 08:04:01.44151101 +0000 UTC m=+1302.261951624" watchObservedRunningTime="2025-12-27 08:04:01.450050195 +0000 UTC m=+1302.270490789" Dec 27 08:04:02 crc kubenswrapper[4934]: I1227 08:04:02.434111 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e59489eb-aaa1-4188-b593-d3d3c2f1f412","Type":"ContainerStarted","Data":"5c5b6b462581059775ce4d6f2188d10e4eff04a5ffac4ed2ddae1aaeff190c06"} Dec 27 08:04:03 crc kubenswrapper[4934]: I1227 08:04:03.461201 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e59489eb-aaa1-4188-b593-d3d3c2f1f412","Type":"ContainerStarted","Data":"fa11fd22429b681a41ccf4e46836333e77a1177cbeeefa67f38aebe703aa3cce"} Dec 27 08:04:04 crc kubenswrapper[4934]: I1227 08:04:04.480435 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e59489eb-aaa1-4188-b593-d3d3c2f1f412","Type":"ContainerStarted","Data":"e78668e3ca47806c9bfb88010d86e84278bb4577bced1680642a5fd6cad9a28e"} Dec 27 08:04:04 crc kubenswrapper[4934]: I1227 08:04:04.481169 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e59489eb-aaa1-4188-b593-d3d3c2f1f412","Type":"ContainerStarted","Data":"75536161f4aded9c4e7c6159973679082ce44cecb69c6fa818e01ae2bfbf66b9"} Dec 27 08:04:04 crc kubenswrapper[4934]: I1227 08:04:04.484700 4934 generic.go:334] "Generic (PLEG): container finished" podID="5c811bf9-1bb4-4f9b-8967-9a67bb94c520" containerID="cf486a099c077f6365252affac7ca7a4613758c7406362dc41872cda5fa53012" exitCode=0 Dec 27 08:04:04 crc kubenswrapper[4934]: I1227 08:04:04.484751 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-cgptv" event={"ID":"5c811bf9-1bb4-4f9b-8967-9a67bb94c520","Type":"ContainerDied","Data":"cf486a099c077f6365252affac7ca7a4613758c7406362dc41872cda5fa53012"} Dec 27 08:04:05 crc kubenswrapper[4934]: I1227 08:04:05.500902 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e59489eb-aaa1-4188-b593-d3d3c2f1f412","Type":"ContainerStarted","Data":"a7c7b5363cfd67d9929a5276f7a1945d899152e8c461d06819c2f6ecee110c18"} Dec 27 08:04:05 crc kubenswrapper[4934]: I1227 08:04:05.501316 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e59489eb-aaa1-4188-b593-d3d3c2f1f412","Type":"ContainerStarted","Data":"83e0dbbd31be62ab142bdd6f22f94d848a0409db918b209c2957dc7d36f4114e"} Dec 27 08:04:05 crc kubenswrapper[4934]: I1227 08:04:05.879595 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-cgptv" Dec 27 08:04:05 crc kubenswrapper[4934]: I1227 08:04:05.987062 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c811bf9-1bb4-4f9b-8967-9a67bb94c520-config-data\") pod \"5c811bf9-1bb4-4f9b-8967-9a67bb94c520\" (UID: \"5c811bf9-1bb4-4f9b-8967-9a67bb94c520\") " Dec 27 08:04:05 crc kubenswrapper[4934]: I1227 08:04:05.987475 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qs64\" (UniqueName: \"kubernetes.io/projected/5c811bf9-1bb4-4f9b-8967-9a67bb94c520-kube-api-access-5qs64\") pod \"5c811bf9-1bb4-4f9b-8967-9a67bb94c520\" (UID: \"5c811bf9-1bb4-4f9b-8967-9a67bb94c520\") " Dec 27 08:04:05 crc kubenswrapper[4934]: I1227 08:04:05.987555 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c811bf9-1bb4-4f9b-8967-9a67bb94c520-combined-ca-bundle\") pod \"5c811bf9-1bb4-4f9b-8967-9a67bb94c520\" (UID: \"5c811bf9-1bb4-4f9b-8967-9a67bb94c520\") " Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.011228 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c811bf9-1bb4-4f9b-8967-9a67bb94c520-kube-api-access-5qs64" (OuterVolumeSpecName: "kube-api-access-5qs64") pod "5c811bf9-1bb4-4f9b-8967-9a67bb94c520" (UID: "5c811bf9-1bb4-4f9b-8967-9a67bb94c520"). InnerVolumeSpecName "kube-api-access-5qs64". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.031538 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c811bf9-1bb4-4f9b-8967-9a67bb94c520-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c811bf9-1bb4-4f9b-8967-9a67bb94c520" (UID: "5c811bf9-1bb4-4f9b-8967-9a67bb94c520"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.050011 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c811bf9-1bb4-4f9b-8967-9a67bb94c520-config-data" (OuterVolumeSpecName: "config-data") pod "5c811bf9-1bb4-4f9b-8967-9a67bb94c520" (UID: "5c811bf9-1bb4-4f9b-8967-9a67bb94c520"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.090539 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qs64\" (UniqueName: \"kubernetes.io/projected/5c811bf9-1bb4-4f9b-8967-9a67bb94c520-kube-api-access-5qs64\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.090569 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c811bf9-1bb4-4f9b-8967-9a67bb94c520-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.090578 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c811bf9-1bb4-4f9b-8967-9a67bb94c520-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.534792 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e59489eb-aaa1-4188-b593-d3d3c2f1f412","Type":"ContainerStarted","Data":"01f50cdc736c60e7a3199357a58cc4fd5a93ba3031eb5678d29859e6d7c79441"} Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.538494 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-cgptv" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.538958 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-cgptv" event={"ID":"5c811bf9-1bb4-4f9b-8967-9a67bb94c520","Type":"ContainerDied","Data":"67260a88d2dca68520e8233eac025f97a21ce3ac8129d3e841f4f7a3b4bce8d5"} Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.539018 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67260a88d2dca68520e8233eac025f97a21ce3ac8129d3e841f4f7a3b4bce8d5" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.540841 4934 generic.go:334] "Generic (PLEG): container finished" podID="e26e3510-3438-4cb9-8d00-f2d0fdf33ef5" containerID="a6882bd1ae74e1d249c2ec243a06cfe7ba98da11d0e0483844d568496945b12a" exitCode=0 Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.540882 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5","Type":"ContainerDied","Data":"a6882bd1ae74e1d249c2ec243a06cfe7ba98da11d0e0483844d568496945b12a"} Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.779223 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-cj7ht"] Dec 27 08:04:06 crc kubenswrapper[4934]: E1227 08:04:06.780126 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c811bf9-1bb4-4f9b-8967-9a67bb94c520" containerName="keystone-db-sync" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.780145 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c811bf9-1bb4-4f9b-8967-9a67bb94c520" containerName="keystone-db-sync" Dec 27 08:04:06 crc kubenswrapper[4934]: E1227 08:04:06.780163 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47797090-7e7d-40ba-bf3f-a43d18af9283" containerName="mariadb-database-create" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.780171 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="47797090-7e7d-40ba-bf3f-a43d18af9283" containerName="mariadb-database-create" Dec 27 08:04:06 crc kubenswrapper[4934]: E1227 08:04:06.780199 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4aa74e5-8fd8-45e5-907d-e95f8bf11d37" containerName="mariadb-database-create" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.780207 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4aa74e5-8fd8-45e5-907d-e95f8bf11d37" containerName="mariadb-database-create" Dec 27 08:04:06 crc kubenswrapper[4934]: E1227 08:04:06.780233 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d745a073-8cd4-463c-ae78-53185e910777" containerName="mariadb-database-create" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.780241 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d745a073-8cd4-463c-ae78-53185e910777" containerName="mariadb-database-create" Dec 27 08:04:06 crc kubenswrapper[4934]: E1227 08:04:06.780265 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06fd3b33-8755-4e16-93df-6ca70836a8f7" containerName="mariadb-database-create" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.780272 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="06fd3b33-8755-4e16-93df-6ca70836a8f7" containerName="mariadb-database-create" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.780556 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c811bf9-1bb4-4f9b-8967-9a67bb94c520" containerName="keystone-db-sync" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.780581 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="d745a073-8cd4-463c-ae78-53185e910777" containerName="mariadb-database-create" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.780603 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="06fd3b33-8755-4e16-93df-6ca70836a8f7" containerName="mariadb-database-create" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.780623 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4aa74e5-8fd8-45e5-907d-e95f8bf11d37" containerName="mariadb-database-create" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.780640 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="47797090-7e7d-40ba-bf3f-a43d18af9283" containerName="mariadb-database-create" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.782564 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.801459 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-cj7ht"] Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.824943 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stbbf\" (UniqueName: \"kubernetes.io/projected/f43cf9bd-c1be-499b-a263-0b7de7de5961-kube-api-access-stbbf\") pod \"dnsmasq-dns-5c9d85d47c-cj7ht\" (UID: \"f43cf9bd-c1be-499b-a263-0b7de7de5961\") " pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.825035 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-config\") pod \"dnsmasq-dns-5c9d85d47c-cj7ht\" (UID: \"f43cf9bd-c1be-499b-a263-0b7de7de5961\") " pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.825055 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9d85d47c-cj7ht\" (UID: \"f43cf9bd-c1be-499b-a263-0b7de7de5961\") " pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.825119 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9d85d47c-cj7ht\" (UID: \"f43cf9bd-c1be-499b-a263-0b7de7de5961\") " pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.825137 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-dns-svc\") pod \"dnsmasq-dns-5c9d85d47c-cj7ht\" (UID: \"f43cf9bd-c1be-499b-a263-0b7de7de5961\") " pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.826506 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-9wq69"] Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.827898 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.832657 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.833109 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.833314 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-8r767" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.834492 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.879308 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9wq69"] Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.926549 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-config\") pod \"dnsmasq-dns-5c9d85d47c-cj7ht\" (UID: \"f43cf9bd-c1be-499b-a263-0b7de7de5961\") " pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.926587 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9d85d47c-cj7ht\" (UID: \"f43cf9bd-c1be-499b-a263-0b7de7de5961\") " pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.926649 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9d85d47c-cj7ht\" (UID: \"f43cf9bd-c1be-499b-a263-0b7de7de5961\") " pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.926669 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-dns-svc\") pod \"dnsmasq-dns-5c9d85d47c-cj7ht\" (UID: \"f43cf9bd-c1be-499b-a263-0b7de7de5961\") " pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.926740 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stbbf\" (UniqueName: \"kubernetes.io/projected/f43cf9bd-c1be-499b-a263-0b7de7de5961-kube-api-access-stbbf\") pod \"dnsmasq-dns-5c9d85d47c-cj7ht\" (UID: \"f43cf9bd-c1be-499b-a263-0b7de7de5961\") " pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.928068 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9d85d47c-cj7ht\" (UID: \"f43cf9bd-c1be-499b-a263-0b7de7de5961\") " pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.928594 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-dns-svc\") pod \"dnsmasq-dns-5c9d85d47c-cj7ht\" (UID: \"f43cf9bd-c1be-499b-a263-0b7de7de5961\") " pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.929898 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-config\") pod \"dnsmasq-dns-5c9d85d47c-cj7ht\" (UID: \"f43cf9bd-c1be-499b-a263-0b7de7de5961\") " pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.932544 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9d85d47c-cj7ht\" (UID: \"f43cf9bd-c1be-499b-a263-0b7de7de5961\") " pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" Dec 27 08:04:06 crc kubenswrapper[4934]: I1227 08:04:06.988347 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stbbf\" (UniqueName: \"kubernetes.io/projected/f43cf9bd-c1be-499b-a263-0b7de7de5961-kube-api-access-stbbf\") pod \"dnsmasq-dns-5c9d85d47c-cj7ht\" (UID: \"f43cf9bd-c1be-499b-a263-0b7de7de5961\") " pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.032662 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-credential-keys\") pod \"keystone-bootstrap-9wq69\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.033122 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvs9m\" (UniqueName: \"kubernetes.io/projected/36677fa3-dfde-47b2-a5bb-524f18a3bc91-kube-api-access-jvs9m\") pod \"keystone-bootstrap-9wq69\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.033152 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-combined-ca-bundle\") pod \"keystone-bootstrap-9wq69\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.033263 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-scripts\") pod \"keystone-bootstrap-9wq69\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.033282 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-config-data\") pod \"keystone-bootstrap-9wq69\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.033375 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-fernet-keys\") pod \"keystone-bootstrap-9wq69\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.073746 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-cj7ht"] Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.074642 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.117924 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-vkcbn"] Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.133065 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-vkcbn" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.141110 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvs9m\" (UniqueName: \"kubernetes.io/projected/36677fa3-dfde-47b2-a5bb-524f18a3bc91-kube-api-access-jvs9m\") pod \"keystone-bootstrap-9wq69\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.141171 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-combined-ca-bundle\") pod \"keystone-bootstrap-9wq69\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.141232 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-scripts\") pod \"keystone-bootstrap-9wq69\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.141252 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-config-data\") pod \"keystone-bootstrap-9wq69\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.141317 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-fernet-keys\") pod \"keystone-bootstrap-9wq69\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.141363 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-credential-keys\") pod \"keystone-bootstrap-9wq69\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.165215 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-credential-keys\") pod \"keystone-bootstrap-9wq69\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.172490 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-scripts\") pod \"keystone-bootstrap-9wq69\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.181579 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-rldj8" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.181840 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.181988 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.186186 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-vkcbn"] Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.207707 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-fernet-keys\") pod \"keystone-bootstrap-9wq69\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.208274 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-combined-ca-bundle\") pod \"keystone-bootstrap-9wq69\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.212683 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-config-data\") pod \"keystone-bootstrap-9wq69\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.236552 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-cm4wh"] Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.249395 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.253871 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-config-data\") pod \"placement-db-sync-vkcbn\" (UID: \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\") " pod="openstack/placement-db-sync-vkcbn" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.253945 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-logs\") pod \"placement-db-sync-vkcbn\" (UID: \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\") " pod="openstack/placement-db-sync-vkcbn" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.253967 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-scripts\") pod \"placement-db-sync-vkcbn\" (UID: \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\") " pod="openstack/placement-db-sync-vkcbn" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.253984 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-combined-ca-bundle\") pod \"placement-db-sync-vkcbn\" (UID: \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\") " pod="openstack/placement-db-sync-vkcbn" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.254229 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9wfl\" (UniqueName: \"kubernetes.io/projected/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-kube-api-access-s9wfl\") pod \"placement-db-sync-vkcbn\" (UID: \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\") " pod="openstack/placement-db-sync-vkcbn" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.260169 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvs9m\" (UniqueName: \"kubernetes.io/projected/36677fa3-dfde-47b2-a5bb-524f18a3bc91-kube-api-access-jvs9m\") pod \"keystone-bootstrap-9wq69\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.281476 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-cm4wh"] Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.327880 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.338630 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.342064 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.346955 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.350846 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.356068 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-config\") pod \"dnsmasq-dns-6ffb94d8ff-cm4wh\" (UID: \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.356122 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-config-data\") pod \"placement-db-sync-vkcbn\" (UID: \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\") " pod="openstack/placement-db-sync-vkcbn" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.356147 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-logs\") pod \"placement-db-sync-vkcbn\" (UID: \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\") " pod="openstack/placement-db-sync-vkcbn" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.356162 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-scripts\") pod \"placement-db-sync-vkcbn\" (UID: \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\") " pod="openstack/placement-db-sync-vkcbn" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.356178 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-combined-ca-bundle\") pod \"placement-db-sync-vkcbn\" (UID: \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\") " pod="openstack/placement-db-sync-vkcbn" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.356221 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prxq6\" (UniqueName: \"kubernetes.io/projected/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-kube-api-access-prxq6\") pod \"dnsmasq-dns-6ffb94d8ff-cm4wh\" (UID: \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.356241 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-ovsdbserver-nb\") pod \"dnsmasq-dns-6ffb94d8ff-cm4wh\" (UID: \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.356277 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-dns-svc\") pod \"dnsmasq-dns-6ffb94d8ff-cm4wh\" (UID: \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.356301 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9wfl\" (UniqueName: \"kubernetes.io/projected/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-kube-api-access-s9wfl\") pod \"placement-db-sync-vkcbn\" (UID: \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\") " pod="openstack/placement-db-sync-vkcbn" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.356355 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-ovsdbserver-sb\") pod \"dnsmasq-dns-6ffb94d8ff-cm4wh\" (UID: \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.357583 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-logs\") pod \"placement-db-sync-vkcbn\" (UID: \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\") " pod="openstack/placement-db-sync-vkcbn" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.361370 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-scripts\") pod \"placement-db-sync-vkcbn\" (UID: \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\") " pod="openstack/placement-db-sync-vkcbn" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.366567 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-combined-ca-bundle\") pod \"placement-db-sync-vkcbn\" (UID: \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\") " pod="openstack/placement-db-sync-vkcbn" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.373159 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-config-data\") pod \"placement-db-sync-vkcbn\" (UID: \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\") " pod="openstack/placement-db-sync-vkcbn" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.393736 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9wfl\" (UniqueName: \"kubernetes.io/projected/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-kube-api-access-s9wfl\") pod \"placement-db-sync-vkcbn\" (UID: \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\") " pod="openstack/placement-db-sync-vkcbn" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.462316 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-dns-svc\") pod \"dnsmasq-dns-6ffb94d8ff-cm4wh\" (UID: \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.462609 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/badf697c-e657-4b80-9483-63c94c87afaf-log-httpd\") pod \"ceilometer-0\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.462700 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/badf697c-e657-4b80-9483-63c94c87afaf-run-httpd\") pod \"ceilometer-0\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.462808 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-ovsdbserver-sb\") pod \"dnsmasq-dns-6ffb94d8ff-cm4wh\" (UID: \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.462930 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.463002 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-config-data\") pod \"ceilometer-0\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.463069 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-scripts\") pod \"ceilometer-0\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.463183 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-config\") pod \"dnsmasq-dns-6ffb94d8ff-cm4wh\" (UID: \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.463283 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.463383 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prxq6\" (UniqueName: \"kubernetes.io/projected/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-kube-api-access-prxq6\") pod \"dnsmasq-dns-6ffb94d8ff-cm4wh\" (UID: \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.463453 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-ovsdbserver-nb\") pod \"dnsmasq-dns-6ffb94d8ff-cm4wh\" (UID: \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.463536 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vksgf\" (UniqueName: \"kubernetes.io/projected/badf697c-e657-4b80-9483-63c94c87afaf-kube-api-access-vksgf\") pod \"ceilometer-0\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.464736 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-dns-svc\") pod \"dnsmasq-dns-6ffb94d8ff-cm4wh\" (UID: \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.464837 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-ovsdbserver-sb\") pod \"dnsmasq-dns-6ffb94d8ff-cm4wh\" (UID: \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.465428 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-config\") pod \"dnsmasq-dns-6ffb94d8ff-cm4wh\" (UID: \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.465661 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-ovsdbserver-nb\") pod \"dnsmasq-dns-6ffb94d8ff-cm4wh\" (UID: \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.478673 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.506550 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-vkcbn" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.516805 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prxq6\" (UniqueName: \"kubernetes.io/projected/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-kube-api-access-prxq6\") pod \"dnsmasq-dns-6ffb94d8ff-cm4wh\" (UID: \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\") " pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.575273 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.575344 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vksgf\" (UniqueName: \"kubernetes.io/projected/badf697c-e657-4b80-9483-63c94c87afaf-kube-api-access-vksgf\") pod \"ceilometer-0\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.575425 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/badf697c-e657-4b80-9483-63c94c87afaf-log-httpd\") pod \"ceilometer-0\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.575454 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/badf697c-e657-4b80-9483-63c94c87afaf-run-httpd\") pod \"ceilometer-0\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.575553 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.575571 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-config-data\") pod \"ceilometer-0\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.575588 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-scripts\") pod \"ceilometer-0\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.576740 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/badf697c-e657-4b80-9483-63c94c87afaf-log-httpd\") pod \"ceilometer-0\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.577935 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/badf697c-e657-4b80-9483-63c94c87afaf-run-httpd\") pod \"ceilometer-0\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.602734 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.604985 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-config-data\") pod \"ceilometer-0\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.608132 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-scripts\") pod \"ceilometer-0\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.610865 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vksgf\" (UniqueName: \"kubernetes.io/projected/badf697c-e657-4b80-9483-63c94c87afaf-kube-api-access-vksgf\") pod \"ceilometer-0\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.614553 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.622261 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.636394 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6vlxg" event={"ID":"94526b19-793f-4ec9-9a80-8a7d4f8cab87","Type":"ContainerStarted","Data":"f33426463ede301b4ff1ea7a6f26b88450efde0a625ec9eda1972bc388900f56"} Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.697520 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.743974 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-6vlxg" podStartSLOduration=2.808086528 podStartE2EDuration="35.743954146s" podCreationTimestamp="2025-12-27 08:03:32 +0000 UTC" firstStartedPulling="2025-12-27 08:03:33.36020172 +0000 UTC m=+1274.180642334" lastFinishedPulling="2025-12-27 08:04:06.296069358 +0000 UTC m=+1307.116509952" observedRunningTime="2025-12-27 08:04:07.676578695 +0000 UTC m=+1308.497019299" watchObservedRunningTime="2025-12-27 08:04:07.743954146 +0000 UTC m=+1308.564394730" Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.785117 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e59489eb-aaa1-4188-b593-d3d3c2f1f412","Type":"ContainerStarted","Data":"02ae4b0e6aeb3f37deee68004f3a25ac742bb25dcfee7cf33a187c85524e59b9"} Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.785160 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e59489eb-aaa1-4188-b593-d3d3c2f1f412","Type":"ContainerStarted","Data":"0507594a493dda8a2b5a4c17bb298fa21c6f84807dfb23a4f1d3a699e6a3d630"} Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.799156 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5","Type":"ContainerStarted","Data":"7992c87371844a8dd41218d6d72e190cb7e3a2add9f8328ab2346817870c5781"} Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.815990 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-cj7ht"] Dec 27 08:04:07 crc kubenswrapper[4934]: I1227 08:04:07.860890 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=38.815875126 podStartE2EDuration="1m4.860865978s" podCreationTimestamp="2025-12-27 08:03:03 +0000 UTC" firstStartedPulling="2025-12-27 08:03:37.568034877 +0000 UTC m=+1278.388475471" lastFinishedPulling="2025-12-27 08:04:03.613025729 +0000 UTC m=+1304.433466323" observedRunningTime="2025-12-27 08:04:07.835687066 +0000 UTC m=+1308.656127660" watchObservedRunningTime="2025-12-27 08:04:07.860865978 +0000 UTC m=+1308.681306572" Dec 27 08:04:07 crc kubenswrapper[4934]: W1227 08:04:07.912251 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf43cf9bd_c1be_499b_a263_0b7de7de5961.slice/crio-48dee04feb7757759ccbe4f8979a2678417411144375846749ad1586587c5e7f WatchSource:0}: Error finding container 48dee04feb7757759ccbe4f8979a2678417411144375846749ad1586587c5e7f: Status 404 returned error can't find the container with id 48dee04feb7757759ccbe4f8979a2678417411144375846749ad1586587c5e7f Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.184464 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9wq69"] Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.248391 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-cm4wh"] Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.301899 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-xpk49"] Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.307605 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.309291 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-xpk49"] Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.310277 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.411611 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drqh2\" (UniqueName: \"kubernetes.io/projected/3d97a13d-c9b8-4629-b847-0c578cdb71ff-kube-api-access-drqh2\") pod \"dnsmasq-dns-cf78879c9-xpk49\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.411650 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-xpk49\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.411714 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-xpk49\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.411745 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-config\") pod \"dnsmasq-dns-cf78879c9-xpk49\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.411817 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-xpk49\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.411906 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-dns-svc\") pod \"dnsmasq-dns-cf78879c9-xpk49\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.422731 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-cm4wh"] Dec 27 08:04:08 crc kubenswrapper[4934]: W1227 08:04:08.423558 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02b67eda_bcca_41bb_86b0_75c9d6dfa62d.slice/crio-f350e2eb51fcbc8aab6814d5e1eef137bbe5498e282301aaadef26ebb02d17bf WatchSource:0}: Error finding container f350e2eb51fcbc8aab6814d5e1eef137bbe5498e282301aaadef26ebb02d17bf: Status 404 returned error can't find the container with id f350e2eb51fcbc8aab6814d5e1eef137bbe5498e282301aaadef26ebb02d17bf Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.478921 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-vkcbn"] Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.513252 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-xpk49\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.513361 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-dns-svc\") pod \"dnsmasq-dns-cf78879c9-xpk49\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.513419 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drqh2\" (UniqueName: \"kubernetes.io/projected/3d97a13d-c9b8-4629-b847-0c578cdb71ff-kube-api-access-drqh2\") pod \"dnsmasq-dns-cf78879c9-xpk49\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.513439 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-xpk49\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.513469 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-xpk49\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.513492 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-config\") pod \"dnsmasq-dns-cf78879c9-xpk49\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.514378 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-config\") pod \"dnsmasq-dns-cf78879c9-xpk49\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.514918 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-xpk49\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.515607 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-dns-svc\") pod \"dnsmasq-dns-cf78879c9-xpk49\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.516142 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-xpk49\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.516534 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-xpk49\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.532750 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drqh2\" (UniqueName: \"kubernetes.io/projected/3d97a13d-c9b8-4629-b847-0c578cdb71ff-kube-api-access-drqh2\") pod \"dnsmasq-dns-cf78879c9-xpk49\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.683191 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.694402 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.826458 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9wq69" event={"ID":"36677fa3-dfde-47b2-a5bb-524f18a3bc91","Type":"ContainerStarted","Data":"032cc47dedce6dbbea921dd0ef1607d653ed430218ac0956abc28d42c32b092e"} Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.826814 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9wq69" event={"ID":"36677fa3-dfde-47b2-a5bb-524f18a3bc91","Type":"ContainerStarted","Data":"964232c355198285ac57b9fdde6e022dc1408c1930156bfac0064f1cbe55700c"} Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.829647 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"badf697c-e657-4b80-9483-63c94c87afaf","Type":"ContainerStarted","Data":"27f762146bcb13869b79a8f4b3f2d96e8251bf8b9104e6f529417e628cdd07c6"} Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.832930 4934 generic.go:334] "Generic (PLEG): container finished" podID="02b67eda-bcca-41bb-86b0-75c9d6dfa62d" containerID="22ac83affd9eaacb6488c4a2c9be6ab02ed280b48a6e9cd5e958cf25d7a12bf7" exitCode=0 Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.833003 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" event={"ID":"02b67eda-bcca-41bb-86b0-75c9d6dfa62d","Type":"ContainerDied","Data":"22ac83affd9eaacb6488c4a2c9be6ab02ed280b48a6e9cd5e958cf25d7a12bf7"} Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.833034 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" event={"ID":"02b67eda-bcca-41bb-86b0-75c9d6dfa62d","Type":"ContainerStarted","Data":"f350e2eb51fcbc8aab6814d5e1eef137bbe5498e282301aaadef26ebb02d17bf"} Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.835741 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-vkcbn" event={"ID":"f7e0de58-4866-4ba7-9a80-7bce4eb52b97","Type":"ContainerStarted","Data":"85231d5f4dd5e43d6234efe94f61a710ac0adb3f4cfd038cc5c24fc483c96be2"} Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.858863 4934 generic.go:334] "Generic (PLEG): container finished" podID="f43cf9bd-c1be-499b-a263-0b7de7de5961" containerID="5805e834233bfbef53115169349dbc519f2b820afe8d0bea2e11c01c167f724c" exitCode=0 Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.859159 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" event={"ID":"f43cf9bd-c1be-499b-a263-0b7de7de5961","Type":"ContainerDied","Data":"5805e834233bfbef53115169349dbc519f2b820afe8d0bea2e11c01c167f724c"} Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.859198 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" event={"ID":"f43cf9bd-c1be-499b-a263-0b7de7de5961","Type":"ContainerStarted","Data":"48dee04feb7757759ccbe4f8979a2678417411144375846749ad1586587c5e7f"} Dec 27 08:04:08 crc kubenswrapper[4934]: I1227 08:04:08.867725 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-9wq69" podStartSLOduration=2.867698383 podStartE2EDuration="2.867698383s" podCreationTimestamp="2025-12-27 08:04:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:04:08.853508997 +0000 UTC m=+1309.673949601" watchObservedRunningTime="2025-12-27 08:04:08.867698383 +0000 UTC m=+1309.688138977" Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.243449 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.259549 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-xpk49"] Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.383152 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-e4e3-account-create-ssb4t"] Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.387804 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e4e3-account-create-ssb4t" Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.390580 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.395894 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-e4e3-account-create-ssb4t"] Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.480377 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lwkg\" (UniqueName: \"kubernetes.io/projected/0328afdb-dc17-4fa6-9d22-44dca770b879-kube-api-access-8lwkg\") pod \"cinder-e4e3-account-create-ssb4t\" (UID: \"0328afdb-dc17-4fa6-9d22-44dca770b879\") " pod="openstack/cinder-e4e3-account-create-ssb4t" Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.569143 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-e80b-account-create-fvzrf"] Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.571953 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-e80b-account-create-fvzrf" Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.573933 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.575390 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-e80b-account-create-fvzrf"] Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.585378 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lwkg\" (UniqueName: \"kubernetes.io/projected/0328afdb-dc17-4fa6-9d22-44dca770b879-kube-api-access-8lwkg\") pod \"cinder-e4e3-account-create-ssb4t\" (UID: \"0328afdb-dc17-4fa6-9d22-44dca770b879\") " pod="openstack/cinder-e4e3-account-create-ssb4t" Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.644131 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-e74b-account-create-dwdlx"] Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.645811 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e74b-account-create-dwdlx" Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.647974 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.648515 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lwkg\" (UniqueName: \"kubernetes.io/projected/0328afdb-dc17-4fa6-9d22-44dca770b879-kube-api-access-8lwkg\") pod \"cinder-e4e3-account-create-ssb4t\" (UID: \"0328afdb-dc17-4fa6-9d22-44dca770b879\") " pod="openstack/cinder-e4e3-account-create-ssb4t" Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.653494 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-e74b-account-create-dwdlx"] Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.689471 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj6v8\" (UniqueName: \"kubernetes.io/projected/e3489623-db32-44b7-9201-83a6f7603021-kube-api-access-jj6v8\") pod \"heat-e80b-account-create-fvzrf\" (UID: \"e3489623-db32-44b7-9201-83a6f7603021\") " pod="openstack/heat-e80b-account-create-fvzrf" Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.757594 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e4e3-account-create-ssb4t" Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.790879 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fjbb\" (UniqueName: \"kubernetes.io/projected/2c23056c-4932-4373-9e14-c3f93989eb7f-kube-api-access-7fjbb\") pod \"barbican-e74b-account-create-dwdlx\" (UID: \"2c23056c-4932-4373-9e14-c3f93989eb7f\") " pod="openstack/barbican-e74b-account-create-dwdlx" Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.790966 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj6v8\" (UniqueName: \"kubernetes.io/projected/e3489623-db32-44b7-9201-83a6f7603021-kube-api-access-jj6v8\") pod \"heat-e80b-account-create-fvzrf\" (UID: \"e3489623-db32-44b7-9201-83a6f7603021\") " pod="openstack/heat-e80b-account-create-fvzrf" Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.792900 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-a9fc-account-create-px94p"] Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.794376 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a9fc-account-create-px94p" Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.798568 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.801896 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-a9fc-account-create-px94p"] Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.825893 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj6v8\" (UniqueName: \"kubernetes.io/projected/e3489623-db32-44b7-9201-83a6f7603021-kube-api-access-jj6v8\") pod \"heat-e80b-account-create-fvzrf\" (UID: \"e3489623-db32-44b7-9201-83a6f7603021\") " pod="openstack/heat-e80b-account-create-fvzrf" Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.870928 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" event={"ID":"f43cf9bd-c1be-499b-a263-0b7de7de5961","Type":"ContainerDied","Data":"48dee04feb7757759ccbe4f8979a2678417411144375846749ad1586587c5e7f"} Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.870960 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48dee04feb7757759ccbe4f8979a2678417411144375846749ad1586587c5e7f" Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.872185 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-xpk49" event={"ID":"3d97a13d-c9b8-4629-b847-0c578cdb71ff","Type":"ContainerStarted","Data":"357fe674f6f3090b04c39c3223b0cb51d459cd002734c58d3bbbf3d6ad20e2d2"} Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.893777 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfxqv\" (UniqueName: \"kubernetes.io/projected/1443d960-8241-4008-9a7a-7d32a158dd32-kube-api-access-rfxqv\") pod \"neutron-a9fc-account-create-px94p\" (UID: \"1443d960-8241-4008-9a7a-7d32a158dd32\") " pod="openstack/neutron-a9fc-account-create-px94p" Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.893884 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fjbb\" (UniqueName: \"kubernetes.io/projected/2c23056c-4932-4373-9e14-c3f93989eb7f-kube-api-access-7fjbb\") pod \"barbican-e74b-account-create-dwdlx\" (UID: \"2c23056c-4932-4373-9e14-c3f93989eb7f\") " pod="openstack/barbican-e74b-account-create-dwdlx" Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.898656 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-e80b-account-create-fvzrf" Dec 27 08:04:09 crc kubenswrapper[4934]: I1227 08:04:09.939978 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fjbb\" (UniqueName: \"kubernetes.io/projected/2c23056c-4932-4373-9e14-c3f93989eb7f-kube-api-access-7fjbb\") pod \"barbican-e74b-account-create-dwdlx\" (UID: \"2c23056c-4932-4373-9e14-c3f93989eb7f\") " pod="openstack/barbican-e74b-account-create-dwdlx" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:09.998368 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfxqv\" (UniqueName: \"kubernetes.io/projected/1443d960-8241-4008-9a7a-7d32a158dd32-kube-api-access-rfxqv\") pod \"neutron-a9fc-account-create-px94p\" (UID: \"1443d960-8241-4008-9a7a-7d32a158dd32\") " pod="openstack/neutron-a9fc-account-create-px94p" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.027718 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfxqv\" (UniqueName: \"kubernetes.io/projected/1443d960-8241-4008-9a7a-7d32a158dd32-kube-api-access-rfxqv\") pod \"neutron-a9fc-account-create-px94p\" (UID: \"1443d960-8241-4008-9a7a-7d32a158dd32\") " pod="openstack/neutron-a9fc-account-create-px94p" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.075800 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e74b-account-create-dwdlx" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.108778 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.109923 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a9fc-account-create-px94p" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.140353 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.207255 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prxq6\" (UniqueName: \"kubernetes.io/projected/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-kube-api-access-prxq6\") pod \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\" (UID: \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\") " Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.207347 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-config\") pod \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\" (UID: \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\") " Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.207391 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-dns-svc\") pod \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\" (UID: \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\") " Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.207462 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-config\") pod \"f43cf9bd-c1be-499b-a263-0b7de7de5961\" (UID: \"f43cf9bd-c1be-499b-a263-0b7de7de5961\") " Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.207504 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-ovsdbserver-nb\") pod \"f43cf9bd-c1be-499b-a263-0b7de7de5961\" (UID: \"f43cf9bd-c1be-499b-a263-0b7de7de5961\") " Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.207574 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-ovsdbserver-sb\") pod \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\" (UID: \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\") " Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.207715 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-ovsdbserver-sb\") pod \"f43cf9bd-c1be-499b-a263-0b7de7de5961\" (UID: \"f43cf9bd-c1be-499b-a263-0b7de7de5961\") " Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.207742 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-ovsdbserver-nb\") pod \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\" (UID: \"02b67eda-bcca-41bb-86b0-75c9d6dfa62d\") " Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.207758 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stbbf\" (UniqueName: \"kubernetes.io/projected/f43cf9bd-c1be-499b-a263-0b7de7de5961-kube-api-access-stbbf\") pod \"f43cf9bd-c1be-499b-a263-0b7de7de5961\" (UID: \"f43cf9bd-c1be-499b-a263-0b7de7de5961\") " Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.207779 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-dns-svc\") pod \"f43cf9bd-c1be-499b-a263-0b7de7de5961\" (UID: \"f43cf9bd-c1be-499b-a263-0b7de7de5961\") " Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.225448 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-kube-api-access-prxq6" (OuterVolumeSpecName: "kube-api-access-prxq6") pod "02b67eda-bcca-41bb-86b0-75c9d6dfa62d" (UID: "02b67eda-bcca-41bb-86b0-75c9d6dfa62d"). InnerVolumeSpecName "kube-api-access-prxq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.226168 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f43cf9bd-c1be-499b-a263-0b7de7de5961-kube-api-access-stbbf" (OuterVolumeSpecName: "kube-api-access-stbbf") pod "f43cf9bd-c1be-499b-a263-0b7de7de5961" (UID: "f43cf9bd-c1be-499b-a263-0b7de7de5961"). InnerVolumeSpecName "kube-api-access-stbbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.258319 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f43cf9bd-c1be-499b-a263-0b7de7de5961" (UID: "f43cf9bd-c1be-499b-a263-0b7de7de5961"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.275841 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-config" (OuterVolumeSpecName: "config") pod "f43cf9bd-c1be-499b-a263-0b7de7de5961" (UID: "f43cf9bd-c1be-499b-a263-0b7de7de5961"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.276093 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "02b67eda-bcca-41bb-86b0-75c9d6dfa62d" (UID: "02b67eda-bcca-41bb-86b0-75c9d6dfa62d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.287184 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f43cf9bd-c1be-499b-a263-0b7de7de5961" (UID: "f43cf9bd-c1be-499b-a263-0b7de7de5961"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.289803 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f43cf9bd-c1be-499b-a263-0b7de7de5961" (UID: "f43cf9bd-c1be-499b-a263-0b7de7de5961"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.311646 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "02b67eda-bcca-41bb-86b0-75c9d6dfa62d" (UID: "02b67eda-bcca-41bb-86b0-75c9d6dfa62d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.313303 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.313339 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.313354 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.313366 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.313378 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stbbf\" (UniqueName: \"kubernetes.io/projected/f43cf9bd-c1be-499b-a263-0b7de7de5961-kube-api-access-stbbf\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.313390 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f43cf9bd-c1be-499b-a263-0b7de7de5961-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.313401 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prxq6\" (UniqueName: \"kubernetes.io/projected/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-kube-api-access-prxq6\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.313412 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.354793 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-config" (OuterVolumeSpecName: "config") pod "02b67eda-bcca-41bb-86b0-75c9d6dfa62d" (UID: "02b67eda-bcca-41bb-86b0-75c9d6dfa62d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.363179 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "02b67eda-bcca-41bb-86b0-75c9d6dfa62d" (UID: "02b67eda-bcca-41bb-86b0-75c9d6dfa62d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.404944 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-e80b-account-create-fvzrf"] Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.435856 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.436875 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02b67eda-bcca-41bb-86b0-75c9d6dfa62d-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.621858 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-e4e3-account-create-ssb4t"] Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.886129 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e4e3-account-create-ssb4t" event={"ID":"0328afdb-dc17-4fa6-9d22-44dca770b879","Type":"ContainerStarted","Data":"394bf1fc414bae74c0e033e6e581554e778c68304631b61a977c84dc37719249"} Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.889479 4934 generic.go:334] "Generic (PLEG): container finished" podID="e3489623-db32-44b7-9201-83a6f7603021" containerID="afb07e08e8bedb3662ba63846f90cd9ae80ee3a32b160bb54e44d523509b36b8" exitCode=0 Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.889523 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-e80b-account-create-fvzrf" event={"ID":"e3489623-db32-44b7-9201-83a6f7603021","Type":"ContainerDied","Data":"afb07e08e8bedb3662ba63846f90cd9ae80ee3a32b160bb54e44d523509b36b8"} Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.889687 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-e80b-account-create-fvzrf" event={"ID":"e3489623-db32-44b7-9201-83a6f7603021","Type":"ContainerStarted","Data":"aefab5cbdd24f598743ac7a54328743ae8076a16a370893e5fb86375f91d9d57"} Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.891443 4934 generic.go:334] "Generic (PLEG): container finished" podID="3d97a13d-c9b8-4629-b847-0c578cdb71ff" containerID="e4d8910c5b76c302c878aa5bac21827e5be05c7f33652ac0692b056f2f80d08a" exitCode=0 Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.891505 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-xpk49" event={"ID":"3d97a13d-c9b8-4629-b847-0c578cdb71ff","Type":"ContainerDied","Data":"e4d8910c5b76c302c878aa5bac21827e5be05c7f33652ac0692b056f2f80d08a"} Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.905640 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" event={"ID":"02b67eda-bcca-41bb-86b0-75c9d6dfa62d","Type":"ContainerDied","Data":"f350e2eb51fcbc8aab6814d5e1eef137bbe5498e282301aaadef26ebb02d17bf"} Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.905955 4934 scope.go:117] "RemoveContainer" containerID="22ac83affd9eaacb6488c4a2c9be6ab02ed280b48a6e9cd5e958cf25d7a12bf7" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.906116 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6ffb94d8ff-cm4wh" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.922303 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9d85d47c-cj7ht" Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.926204 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5","Type":"ContainerStarted","Data":"285cc0703f18e650918f35919c201c956fafe0d9c04e79af104db57ae415b1df"} Dec 27 08:04:10 crc kubenswrapper[4934]: I1227 08:04:10.926254 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"e26e3510-3438-4cb9-8d00-f2d0fdf33ef5","Type":"ContainerStarted","Data":"1c90d6b33e2c2ee0796f44cfdbc665f328f6f0aa2a0f54cc5a73a159a8e0dd10"} Dec 27 08:04:11 crc kubenswrapper[4934]: I1227 08:04:11.031807 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=17.031791935 podStartE2EDuration="17.031791935s" podCreationTimestamp="2025-12-27 08:03:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:04:10.984511849 +0000 UTC m=+1311.804952443" watchObservedRunningTime="2025-12-27 08:04:11.031791935 +0000 UTC m=+1311.852232529" Dec 27 08:04:11 crc kubenswrapper[4934]: I1227 08:04:11.049141 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-e74b-account-create-dwdlx"] Dec 27 08:04:11 crc kubenswrapper[4934]: W1227 08:04:11.056688 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1443d960_8241_4008_9a7a_7d32a158dd32.slice/crio-995f2010e1b582b314f1e4bdcc07c34e72b1918bb07b8c594f832772e9267d7d WatchSource:0}: Error finding container 995f2010e1b582b314f1e4bdcc07c34e72b1918bb07b8c594f832772e9267d7d: Status 404 returned error can't find the container with id 995f2010e1b582b314f1e4bdcc07c34e72b1918bb07b8c594f832772e9267d7d Dec 27 08:04:11 crc kubenswrapper[4934]: I1227 08:04:11.061053 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-a9fc-account-create-px94p"] Dec 27 08:04:11 crc kubenswrapper[4934]: I1227 08:04:11.118136 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-cm4wh"] Dec 27 08:04:11 crc kubenswrapper[4934]: I1227 08:04:11.146045 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6ffb94d8ff-cm4wh"] Dec 27 08:04:11 crc kubenswrapper[4934]: I1227 08:04:11.192009 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-cj7ht"] Dec 27 08:04:11 crc kubenswrapper[4934]: I1227 08:04:11.205468 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9d85d47c-cj7ht"] Dec 27 08:04:11 crc kubenswrapper[4934]: I1227 08:04:11.480009 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02b67eda-bcca-41bb-86b0-75c9d6dfa62d" path="/var/lib/kubelet/pods/02b67eda-bcca-41bb-86b0-75c9d6dfa62d/volumes" Dec 27 08:04:11 crc kubenswrapper[4934]: I1227 08:04:11.481062 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f43cf9bd-c1be-499b-a263-0b7de7de5961" path="/var/lib/kubelet/pods/f43cf9bd-c1be-499b-a263-0b7de7de5961/volumes" Dec 27 08:04:11 crc kubenswrapper[4934]: I1227 08:04:11.944391 4934 generic.go:334] "Generic (PLEG): container finished" podID="2c23056c-4932-4373-9e14-c3f93989eb7f" containerID="6edd9147c9e2c957f7c485e8994dc93a28028e64b952d232a8c9a535692b0117" exitCode=0 Dec 27 08:04:11 crc kubenswrapper[4934]: I1227 08:04:11.944491 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e74b-account-create-dwdlx" event={"ID":"2c23056c-4932-4373-9e14-c3f93989eb7f","Type":"ContainerDied","Data":"6edd9147c9e2c957f7c485e8994dc93a28028e64b952d232a8c9a535692b0117"} Dec 27 08:04:11 crc kubenswrapper[4934]: I1227 08:04:11.944525 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e74b-account-create-dwdlx" event={"ID":"2c23056c-4932-4373-9e14-c3f93989eb7f","Type":"ContainerStarted","Data":"878863974f23e4ffc35d5d594debf548765074ad14a28db06d3bc52116c65401"} Dec 27 08:04:11 crc kubenswrapper[4934]: I1227 08:04:11.947609 4934 generic.go:334] "Generic (PLEG): container finished" podID="1443d960-8241-4008-9a7a-7d32a158dd32" containerID="f82dec17909b52ec75d630a98ad35e72ac5164071010cf7cd387118c867dc70d" exitCode=0 Dec 27 08:04:11 crc kubenswrapper[4934]: I1227 08:04:11.947697 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-a9fc-account-create-px94p" event={"ID":"1443d960-8241-4008-9a7a-7d32a158dd32","Type":"ContainerDied","Data":"f82dec17909b52ec75d630a98ad35e72ac5164071010cf7cd387118c867dc70d"} Dec 27 08:04:11 crc kubenswrapper[4934]: I1227 08:04:11.947745 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-a9fc-account-create-px94p" event={"ID":"1443d960-8241-4008-9a7a-7d32a158dd32","Type":"ContainerStarted","Data":"995f2010e1b582b314f1e4bdcc07c34e72b1918bb07b8c594f832772e9267d7d"} Dec 27 08:04:11 crc kubenswrapper[4934]: I1227 08:04:11.949840 4934 generic.go:334] "Generic (PLEG): container finished" podID="0328afdb-dc17-4fa6-9d22-44dca770b879" containerID="ec3f9d839dd4caf8943fdebc447b00e343dfdebc72e74bb103a9ad1cbd5d5875" exitCode=0 Dec 27 08:04:11 crc kubenswrapper[4934]: I1227 08:04:11.949903 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e4e3-account-create-ssb4t" event={"ID":"0328afdb-dc17-4fa6-9d22-44dca770b879","Type":"ContainerDied","Data":"ec3f9d839dd4caf8943fdebc447b00e343dfdebc72e74bb103a9ad1cbd5d5875"} Dec 27 08:04:11 crc kubenswrapper[4934]: I1227 08:04:11.953578 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-xpk49" event={"ID":"3d97a13d-c9b8-4629-b847-0c578cdb71ff","Type":"ContainerStarted","Data":"40bd45abc7cf4b892b27a1e4eec63f76876fbc3deba5ca067b98e910781c04ee"} Dec 27 08:04:11 crc kubenswrapper[4934]: I1227 08:04:11.954010 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:12 crc kubenswrapper[4934]: I1227 08:04:12.010882 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cf78879c9-xpk49" podStartSLOduration=4.010865823 podStartE2EDuration="4.010865823s" podCreationTimestamp="2025-12-27 08:04:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:04:12.006945795 +0000 UTC m=+1312.827386379" watchObservedRunningTime="2025-12-27 08:04:12.010865823 +0000 UTC m=+1312.831306417" Dec 27 08:04:12 crc kubenswrapper[4934]: I1227 08:04:12.389628 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-e80b-account-create-fvzrf" Dec 27 08:04:12 crc kubenswrapper[4934]: I1227 08:04:12.496163 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jj6v8\" (UniqueName: \"kubernetes.io/projected/e3489623-db32-44b7-9201-83a6f7603021-kube-api-access-jj6v8\") pod \"e3489623-db32-44b7-9201-83a6f7603021\" (UID: \"e3489623-db32-44b7-9201-83a6f7603021\") " Dec 27 08:04:12 crc kubenswrapper[4934]: I1227 08:04:12.517351 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3489623-db32-44b7-9201-83a6f7603021-kube-api-access-jj6v8" (OuterVolumeSpecName: "kube-api-access-jj6v8") pod "e3489623-db32-44b7-9201-83a6f7603021" (UID: "e3489623-db32-44b7-9201-83a6f7603021"). InnerVolumeSpecName "kube-api-access-jj6v8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:04:12 crc kubenswrapper[4934]: I1227 08:04:12.599186 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jj6v8\" (UniqueName: \"kubernetes.io/projected/e3489623-db32-44b7-9201-83a6f7603021-kube-api-access-jj6v8\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:12 crc kubenswrapper[4934]: I1227 08:04:12.979523 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-e80b-account-create-fvzrf" Dec 27 08:04:12 crc kubenswrapper[4934]: I1227 08:04:12.982261 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-e80b-account-create-fvzrf" event={"ID":"e3489623-db32-44b7-9201-83a6f7603021","Type":"ContainerDied","Data":"aefab5cbdd24f598743ac7a54328743ae8076a16a370893e5fb86375f91d9d57"} Dec 27 08:04:12 crc kubenswrapper[4934]: I1227 08:04:12.982311 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aefab5cbdd24f598743ac7a54328743ae8076a16a370893e5fb86375f91d9d57" Dec 27 08:04:13 crc kubenswrapper[4934]: I1227 08:04:13.994593 4934 generic.go:334] "Generic (PLEG): container finished" podID="36677fa3-dfde-47b2-a5bb-524f18a3bc91" containerID="032cc47dedce6dbbea921dd0ef1607d653ed430218ac0956abc28d42c32b092e" exitCode=0 Dec 27 08:04:13 crc kubenswrapper[4934]: I1227 08:04:13.994676 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9wq69" event={"ID":"36677fa3-dfde-47b2-a5bb-524f18a3bc91","Type":"ContainerDied","Data":"032cc47dedce6dbbea921dd0ef1607d653ed430218ac0956abc28d42c32b092e"} Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.743328 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.775244 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-5qllw"] Dec 27 08:04:14 crc kubenswrapper[4934]: E1227 08:04:14.775966 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f43cf9bd-c1be-499b-a263-0b7de7de5961" containerName="init" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.776097 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f43cf9bd-c1be-499b-a263-0b7de7de5961" containerName="init" Dec 27 08:04:14 crc kubenswrapper[4934]: E1227 08:04:14.776180 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02b67eda-bcca-41bb-86b0-75c9d6dfa62d" containerName="init" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.776246 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="02b67eda-bcca-41bb-86b0-75c9d6dfa62d" containerName="init" Dec 27 08:04:14 crc kubenswrapper[4934]: E1227 08:04:14.776359 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3489623-db32-44b7-9201-83a6f7603021" containerName="mariadb-account-create" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.776436 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3489623-db32-44b7-9201-83a6f7603021" containerName="mariadb-account-create" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.776800 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="02b67eda-bcca-41bb-86b0-75c9d6dfa62d" containerName="init" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.776899 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3489623-db32-44b7-9201-83a6f7603021" containerName="mariadb-account-create" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.776992 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f43cf9bd-c1be-499b-a263-0b7de7de5961" containerName="init" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.777930 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-5qllw" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.780810 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-7lm5l" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.781026 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.799280 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-5qllw"] Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.858466 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61f5132a-303d-4f76-ae58-c57d82a59fd4-config-data\") pod \"heat-db-sync-5qllw\" (UID: \"61f5132a-303d-4f76-ae58-c57d82a59fd4\") " pod="openstack/heat-db-sync-5qllw" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.858752 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s9pr\" (UniqueName: \"kubernetes.io/projected/61f5132a-303d-4f76-ae58-c57d82a59fd4-kube-api-access-9s9pr\") pod \"heat-db-sync-5qllw\" (UID: \"61f5132a-303d-4f76-ae58-c57d82a59fd4\") " pod="openstack/heat-db-sync-5qllw" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.858920 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61f5132a-303d-4f76-ae58-c57d82a59fd4-combined-ca-bundle\") pod \"heat-db-sync-5qllw\" (UID: \"61f5132a-303d-4f76-ae58-c57d82a59fd4\") " pod="openstack/heat-db-sync-5qllw" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.941863 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e74b-account-create-dwdlx" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.961356 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fjbb\" (UniqueName: \"kubernetes.io/projected/2c23056c-4932-4373-9e14-c3f93989eb7f-kube-api-access-7fjbb\") pod \"2c23056c-4932-4373-9e14-c3f93989eb7f\" (UID: \"2c23056c-4932-4373-9e14-c3f93989eb7f\") " Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.961854 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61f5132a-303d-4f76-ae58-c57d82a59fd4-config-data\") pod \"heat-db-sync-5qllw\" (UID: \"61f5132a-303d-4f76-ae58-c57d82a59fd4\") " pod="openstack/heat-db-sync-5qllw" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.961904 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s9pr\" (UniqueName: \"kubernetes.io/projected/61f5132a-303d-4f76-ae58-c57d82a59fd4-kube-api-access-9s9pr\") pod \"heat-db-sync-5qllw\" (UID: \"61f5132a-303d-4f76-ae58-c57d82a59fd4\") " pod="openstack/heat-db-sync-5qllw" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.962031 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61f5132a-303d-4f76-ae58-c57d82a59fd4-combined-ca-bundle\") pod \"heat-db-sync-5qllw\" (UID: \"61f5132a-303d-4f76-ae58-c57d82a59fd4\") " pod="openstack/heat-db-sync-5qllw" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.968875 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c23056c-4932-4373-9e14-c3f93989eb7f-kube-api-access-7fjbb" (OuterVolumeSpecName: "kube-api-access-7fjbb") pod "2c23056c-4932-4373-9e14-c3f93989eb7f" (UID: "2c23056c-4932-4373-9e14-c3f93989eb7f"). InnerVolumeSpecName "kube-api-access-7fjbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.969511 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61f5132a-303d-4f76-ae58-c57d82a59fd4-config-data\") pod \"heat-db-sync-5qllw\" (UID: \"61f5132a-303d-4f76-ae58-c57d82a59fd4\") " pod="openstack/heat-db-sync-5qllw" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.970596 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61f5132a-303d-4f76-ae58-c57d82a59fd4-combined-ca-bundle\") pod \"heat-db-sync-5qllw\" (UID: \"61f5132a-303d-4f76-ae58-c57d82a59fd4\") " pod="openstack/heat-db-sync-5qllw" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.973175 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a9fc-account-create-px94p" Dec 27 08:04:14 crc kubenswrapper[4934]: I1227 08:04:14.984424 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s9pr\" (UniqueName: \"kubernetes.io/projected/61f5132a-303d-4f76-ae58-c57d82a59fd4-kube-api-access-9s9pr\") pod \"heat-db-sync-5qllw\" (UID: \"61f5132a-303d-4f76-ae58-c57d82a59fd4\") " pod="openstack/heat-db-sync-5qllw" Dec 27 08:04:15 crc kubenswrapper[4934]: I1227 08:04:15.012034 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e4e3-account-create-ssb4t" event={"ID":"0328afdb-dc17-4fa6-9d22-44dca770b879","Type":"ContainerDied","Data":"394bf1fc414bae74c0e033e6e581554e778c68304631b61a977c84dc37719249"} Dec 27 08:04:15 crc kubenswrapper[4934]: I1227 08:04:15.012074 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="394bf1fc414bae74c0e033e6e581554e778c68304631b61a977c84dc37719249" Dec 27 08:04:15 crc kubenswrapper[4934]: I1227 08:04:15.014155 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-e74b-account-create-dwdlx" event={"ID":"2c23056c-4932-4373-9e14-c3f93989eb7f","Type":"ContainerDied","Data":"878863974f23e4ffc35d5d594debf548765074ad14a28db06d3bc52116c65401"} Dec 27 08:04:15 crc kubenswrapper[4934]: I1227 08:04:15.014191 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-e74b-account-create-dwdlx" Dec 27 08:04:15 crc kubenswrapper[4934]: I1227 08:04:15.014195 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="878863974f23e4ffc35d5d594debf548765074ad14a28db06d3bc52116c65401" Dec 27 08:04:15 crc kubenswrapper[4934]: I1227 08:04:15.023416 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a9fc-account-create-px94p" Dec 27 08:04:15 crc kubenswrapper[4934]: I1227 08:04:15.023885 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-a9fc-account-create-px94p" event={"ID":"1443d960-8241-4008-9a7a-7d32a158dd32","Type":"ContainerDied","Data":"995f2010e1b582b314f1e4bdcc07c34e72b1918bb07b8c594f832772e9267d7d"} Dec 27 08:04:15 crc kubenswrapper[4934]: I1227 08:04:15.023912 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="995f2010e1b582b314f1e4bdcc07c34e72b1918bb07b8c594f832772e9267d7d" Dec 27 08:04:15 crc kubenswrapper[4934]: I1227 08:04:15.063247 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfxqv\" (UniqueName: \"kubernetes.io/projected/1443d960-8241-4008-9a7a-7d32a158dd32-kube-api-access-rfxqv\") pod \"1443d960-8241-4008-9a7a-7d32a158dd32\" (UID: \"1443d960-8241-4008-9a7a-7d32a158dd32\") " Dec 27 08:04:15 crc kubenswrapper[4934]: I1227 08:04:15.064255 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fjbb\" (UniqueName: \"kubernetes.io/projected/2c23056c-4932-4373-9e14-c3f93989eb7f-kube-api-access-7fjbb\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:15 crc kubenswrapper[4934]: I1227 08:04:15.066757 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1443d960-8241-4008-9a7a-7d32a158dd32-kube-api-access-rfxqv" (OuterVolumeSpecName: "kube-api-access-rfxqv") pod "1443d960-8241-4008-9a7a-7d32a158dd32" (UID: "1443d960-8241-4008-9a7a-7d32a158dd32"). InnerVolumeSpecName "kube-api-access-rfxqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:04:15 crc kubenswrapper[4934]: I1227 08:04:15.107842 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-5qllw" Dec 27 08:04:15 crc kubenswrapper[4934]: I1227 08:04:15.131519 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e4e3-account-create-ssb4t" Dec 27 08:04:15 crc kubenswrapper[4934]: I1227 08:04:15.165629 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lwkg\" (UniqueName: \"kubernetes.io/projected/0328afdb-dc17-4fa6-9d22-44dca770b879-kube-api-access-8lwkg\") pod \"0328afdb-dc17-4fa6-9d22-44dca770b879\" (UID: \"0328afdb-dc17-4fa6-9d22-44dca770b879\") " Dec 27 08:04:15 crc kubenswrapper[4934]: I1227 08:04:15.166445 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfxqv\" (UniqueName: \"kubernetes.io/projected/1443d960-8241-4008-9a7a-7d32a158dd32-kube-api-access-rfxqv\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:15 crc kubenswrapper[4934]: I1227 08:04:15.169757 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0328afdb-dc17-4fa6-9d22-44dca770b879-kube-api-access-8lwkg" (OuterVolumeSpecName: "kube-api-access-8lwkg") pod "0328afdb-dc17-4fa6-9d22-44dca770b879" (UID: "0328afdb-dc17-4fa6-9d22-44dca770b879"). InnerVolumeSpecName "kube-api-access-8lwkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:04:15 crc kubenswrapper[4934]: I1227 08:04:15.279432 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lwkg\" (UniqueName: \"kubernetes.io/projected/0328afdb-dc17-4fa6-9d22-44dca770b879-kube-api-access-8lwkg\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:15 crc kubenswrapper[4934]: I1227 08:04:15.331773 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:04:15 crc kubenswrapper[4934]: I1227 08:04:15.331860 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:04:16 crc kubenswrapper[4934]: I1227 08:04:16.047828 4934 generic.go:334] "Generic (PLEG): container finished" podID="94526b19-793f-4ec9-9a80-8a7d4f8cab87" containerID="f33426463ede301b4ff1ea7a6f26b88450efde0a625ec9eda1972bc388900f56" exitCode=0 Dec 27 08:04:16 crc kubenswrapper[4934]: I1227 08:04:16.047941 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6vlxg" event={"ID":"94526b19-793f-4ec9-9a80-8a7d4f8cab87","Type":"ContainerDied","Data":"f33426463ede301b4ff1ea7a6f26b88450efde0a625ec9eda1972bc388900f56"} Dec 27 08:04:16 crc kubenswrapper[4934]: I1227 08:04:16.048183 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e4e3-account-create-ssb4t" Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.799298 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.823242 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6vlxg" Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.846894 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rt5pp\" (UniqueName: \"kubernetes.io/projected/94526b19-793f-4ec9-9a80-8a7d4f8cab87-kube-api-access-rt5pp\") pod \"94526b19-793f-4ec9-9a80-8a7d4f8cab87\" (UID: \"94526b19-793f-4ec9-9a80-8a7d4f8cab87\") " Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.846998 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvs9m\" (UniqueName: \"kubernetes.io/projected/36677fa3-dfde-47b2-a5bb-524f18a3bc91-kube-api-access-jvs9m\") pod \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.847198 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-config-data\") pod \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.847245 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-credential-keys\") pod \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.847265 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94526b19-793f-4ec9-9a80-8a7d4f8cab87-combined-ca-bundle\") pod \"94526b19-793f-4ec9-9a80-8a7d4f8cab87\" (UID: \"94526b19-793f-4ec9-9a80-8a7d4f8cab87\") " Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.847283 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-fernet-keys\") pod \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.847367 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94526b19-793f-4ec9-9a80-8a7d4f8cab87-config-data\") pod \"94526b19-793f-4ec9-9a80-8a7d4f8cab87\" (UID: \"94526b19-793f-4ec9-9a80-8a7d4f8cab87\") " Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.847385 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-scripts\") pod \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.847527 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/94526b19-793f-4ec9-9a80-8a7d4f8cab87-db-sync-config-data\") pod \"94526b19-793f-4ec9-9a80-8a7d4f8cab87\" (UID: \"94526b19-793f-4ec9-9a80-8a7d4f8cab87\") " Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.847569 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-combined-ca-bundle\") pod \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\" (UID: \"36677fa3-dfde-47b2-a5bb-524f18a3bc91\") " Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.854471 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-scripts" (OuterVolumeSpecName: "scripts") pod "36677fa3-dfde-47b2-a5bb-524f18a3bc91" (UID: "36677fa3-dfde-47b2-a5bb-524f18a3bc91"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.854694 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36677fa3-dfde-47b2-a5bb-524f18a3bc91-kube-api-access-jvs9m" (OuterVolumeSpecName: "kube-api-access-jvs9m") pod "36677fa3-dfde-47b2-a5bb-524f18a3bc91" (UID: "36677fa3-dfde-47b2-a5bb-524f18a3bc91"). InnerVolumeSpecName "kube-api-access-jvs9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.855630 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94526b19-793f-4ec9-9a80-8a7d4f8cab87-kube-api-access-rt5pp" (OuterVolumeSpecName: "kube-api-access-rt5pp") pod "94526b19-793f-4ec9-9a80-8a7d4f8cab87" (UID: "94526b19-793f-4ec9-9a80-8a7d4f8cab87"). InnerVolumeSpecName "kube-api-access-rt5pp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.857502 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94526b19-793f-4ec9-9a80-8a7d4f8cab87-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "94526b19-793f-4ec9-9a80-8a7d4f8cab87" (UID: "94526b19-793f-4ec9-9a80-8a7d4f8cab87"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.869309 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "36677fa3-dfde-47b2-a5bb-524f18a3bc91" (UID: "36677fa3-dfde-47b2-a5bb-524f18a3bc91"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.873252 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "36677fa3-dfde-47b2-a5bb-524f18a3bc91" (UID: "36677fa3-dfde-47b2-a5bb-524f18a3bc91"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.895245 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94526b19-793f-4ec9-9a80-8a7d4f8cab87-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94526b19-793f-4ec9-9a80-8a7d4f8cab87" (UID: "94526b19-793f-4ec9-9a80-8a7d4f8cab87"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.897994 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36677fa3-dfde-47b2-a5bb-524f18a3bc91" (UID: "36677fa3-dfde-47b2-a5bb-524f18a3bc91"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.900229 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-config-data" (OuterVolumeSpecName: "config-data") pod "36677fa3-dfde-47b2-a5bb-524f18a3bc91" (UID: "36677fa3-dfde-47b2-a5bb-524f18a3bc91"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.900892 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-5qllw"] Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.950385 4934 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/94526b19-793f-4ec9-9a80-8a7d4f8cab87-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.950423 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.950436 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rt5pp\" (UniqueName: \"kubernetes.io/projected/94526b19-793f-4ec9-9a80-8a7d4f8cab87-kube-api-access-rt5pp\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.950452 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvs9m\" (UniqueName: \"kubernetes.io/projected/36677fa3-dfde-47b2-a5bb-524f18a3bc91-kube-api-access-jvs9m\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.950465 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.950474 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94526b19-793f-4ec9-9a80-8a7d4f8cab87-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.950484 4934 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.950494 4934 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.950503 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36677fa3-dfde-47b2-a5bb-524f18a3bc91-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:17 crc kubenswrapper[4934]: I1227 08:04:17.960225 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94526b19-793f-4ec9-9a80-8a7d4f8cab87-config-data" (OuterVolumeSpecName: "config-data") pod "94526b19-793f-4ec9-9a80-8a7d4f8cab87" (UID: "94526b19-793f-4ec9-9a80-8a7d4f8cab87"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.052807 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94526b19-793f-4ec9-9a80-8a7d4f8cab87-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.068671 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9wq69" event={"ID":"36677fa3-dfde-47b2-a5bb-524f18a3bc91","Type":"ContainerDied","Data":"964232c355198285ac57b9fdde6e022dc1408c1930156bfac0064f1cbe55700c"} Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.068925 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="964232c355198285ac57b9fdde6e022dc1408c1930156bfac0064f1cbe55700c" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.068725 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9wq69" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.069692 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-5qllw" event={"ID":"61f5132a-303d-4f76-ae58-c57d82a59fd4","Type":"ContainerStarted","Data":"75b225b078c8d1ec9d82022861232f98fd169bb3b1758cfaac38f3787e83f71a"} Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.071646 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"badf697c-e657-4b80-9483-63c94c87afaf","Type":"ContainerStarted","Data":"45a559573157263c959004566fc36f27391dce93ff6361826ad2d9438ebd9a66"} Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.073440 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-6vlxg" event={"ID":"94526b19-793f-4ec9-9a80-8a7d4f8cab87","Type":"ContainerDied","Data":"5ab7f8abc9597cb13455904036db82aad1a788a11cb993a7a718c8afa4b94b2f"} Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.073469 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ab7f8abc9597cb13455904036db82aad1a788a11cb993a7a718c8afa4b94b2f" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.073540 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-6vlxg" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.075657 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-vkcbn" event={"ID":"f7e0de58-4866-4ba7-9a80-7bce4eb52b97","Type":"ContainerStarted","Data":"c3f86223e2285fd2790d23cc1157f81d5862840766502576d5b20ca6aac68bd2"} Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.111105 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-vkcbn" podStartSLOduration=2.172022375 podStartE2EDuration="11.111063295s" podCreationTimestamp="2025-12-27 08:04:07 +0000 UTC" firstStartedPulling="2025-12-27 08:04:08.477930516 +0000 UTC m=+1309.298371110" lastFinishedPulling="2025-12-27 08:04:17.416971436 +0000 UTC m=+1318.237412030" observedRunningTime="2025-12-27 08:04:18.096564262 +0000 UTC m=+1318.917004876" watchObservedRunningTime="2025-12-27 08:04:18.111063295 +0000 UTC m=+1318.931503889" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.448649 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-xpk49"] Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.449269 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cf78879c9-xpk49" podUID="3d97a13d-c9b8-4629-b847-0c578cdb71ff" containerName="dnsmasq-dns" containerID="cri-o://40bd45abc7cf4b892b27a1e4eec63f76876fbc3deba5ca067b98e910781c04ee" gracePeriod=10 Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.461445 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.467364 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-dw8tc"] Dec 27 08:04:18 crc kubenswrapper[4934]: E1227 08:04:18.469510 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c23056c-4932-4373-9e14-c3f93989eb7f" containerName="mariadb-account-create" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.469529 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c23056c-4932-4373-9e14-c3f93989eb7f" containerName="mariadb-account-create" Dec 27 08:04:18 crc kubenswrapper[4934]: E1227 08:04:18.469546 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36677fa3-dfde-47b2-a5bb-524f18a3bc91" containerName="keystone-bootstrap" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.469552 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="36677fa3-dfde-47b2-a5bb-524f18a3bc91" containerName="keystone-bootstrap" Dec 27 08:04:18 crc kubenswrapper[4934]: E1227 08:04:18.469575 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94526b19-793f-4ec9-9a80-8a7d4f8cab87" containerName="glance-db-sync" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.469581 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="94526b19-793f-4ec9-9a80-8a7d4f8cab87" containerName="glance-db-sync" Dec 27 08:04:18 crc kubenswrapper[4934]: E1227 08:04:18.469599 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1443d960-8241-4008-9a7a-7d32a158dd32" containerName="mariadb-account-create" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.469605 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1443d960-8241-4008-9a7a-7d32a158dd32" containerName="mariadb-account-create" Dec 27 08:04:18 crc kubenswrapper[4934]: E1227 08:04:18.470171 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0328afdb-dc17-4fa6-9d22-44dca770b879" containerName="mariadb-account-create" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.470179 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="0328afdb-dc17-4fa6-9d22-44dca770b879" containerName="mariadb-account-create" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.470403 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="36677fa3-dfde-47b2-a5bb-524f18a3bc91" containerName="keystone-bootstrap" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.470426 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="0328afdb-dc17-4fa6-9d22-44dca770b879" containerName="mariadb-account-create" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.470435 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="94526b19-793f-4ec9-9a80-8a7d4f8cab87" containerName="glance-db-sync" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.470584 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1443d960-8241-4008-9a7a-7d32a158dd32" containerName="mariadb-account-create" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.470602 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c23056c-4932-4373-9e14-c3f93989eb7f" containerName="mariadb-account-create" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.472470 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.570234 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-dw8tc"] Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.670576 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-dw8tc\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.670635 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-dw8tc\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.670654 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-dw8tc\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.670779 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-config\") pod \"dnsmasq-dns-56df8fb6b7-dw8tc\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.670800 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-dw8tc\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.670842 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-657bv\" (UniqueName: \"kubernetes.io/projected/1df64eee-76a8-4be2-8b36-665bc0f1564f-kube-api-access-657bv\") pod \"dnsmasq-dns-56df8fb6b7-dw8tc\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.684837 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-cf78879c9-xpk49" podUID="3d97a13d-c9b8-4629-b847-0c578cdb71ff" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.177:5353: connect: connection refused" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.773174 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-dw8tc\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.773221 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-dw8tc\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.773241 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-dw8tc\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.773352 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-config\") pod \"dnsmasq-dns-56df8fb6b7-dw8tc\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.773378 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-dw8tc\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.773418 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-657bv\" (UniqueName: \"kubernetes.io/projected/1df64eee-76a8-4be2-8b36-665bc0f1564f-kube-api-access-657bv\") pod \"dnsmasq-dns-56df8fb6b7-dw8tc\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.774169 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-dw8tc\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.774177 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-dw8tc\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.774393 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-dw8tc\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.774403 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-config\") pod \"dnsmasq-dns-56df8fb6b7-dw8tc\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.775064 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-dw8tc\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.790597 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-657bv\" (UniqueName: \"kubernetes.io/projected/1df64eee-76a8-4be2-8b36-665bc0f1564f-kube-api-access-657bv\") pod \"dnsmasq-dns-56df8fb6b7-dw8tc\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:18 crc kubenswrapper[4934]: I1227 08:04:18.805801 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.009069 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-9wq69"] Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.027641 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-9wq69"] Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.130692 4934 generic.go:334] "Generic (PLEG): container finished" podID="3d97a13d-c9b8-4629-b847-0c578cdb71ff" containerID="40bd45abc7cf4b892b27a1e4eec63f76876fbc3deba5ca067b98e910781c04ee" exitCode=0 Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.131166 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-xpk49" event={"ID":"3d97a13d-c9b8-4629-b847-0c578cdb71ff","Type":"ContainerDied","Data":"40bd45abc7cf4b892b27a1e4eec63f76876fbc3deba5ca067b98e910781c04ee"} Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.141435 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-v7n52"] Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.142880 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.153665 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.153697 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-8r767" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.153880 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.153892 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.172987 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-v7n52"] Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.298071 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-fernet-keys\") pod \"keystone-bootstrap-v7n52\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.298269 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-credential-keys\") pod \"keystone-bootstrap-v7n52\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.298337 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-scripts\") pod \"keystone-bootstrap-v7n52\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.305471 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-combined-ca-bundle\") pod \"keystone-bootstrap-v7n52\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.305525 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-config-data\") pod \"keystone-bootstrap-v7n52\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.305617 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9fd6\" (UniqueName: \"kubernetes.io/projected/77f95c7a-ce32-4353-8d7f-c5710d0b0057-kube-api-access-d9fd6\") pod \"keystone-bootstrap-v7n52\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.305740 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.309613 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.313522 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-9x8k8" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.313679 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.313789 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.323193 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.407528 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfm7p\" (UniqueName: \"kubernetes.io/projected/e22ff903-0c33-4be3-9905-d65afb2e6d22-kube-api-access-lfm7p\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.407584 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9fd6\" (UniqueName: \"kubernetes.io/projected/77f95c7a-ce32-4353-8d7f-c5710d0b0057-kube-api-access-d9fd6\") pod \"keystone-bootstrap-v7n52\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.407613 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e22ff903-0c33-4be3-9905-d65afb2e6d22-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.407636 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e22ff903-0c33-4be3-9905-d65afb2e6d22-logs\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.407656 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e22ff903-0c33-4be3-9905-d65afb2e6d22-scripts\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.407698 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-fernet-keys\") pod \"keystone-bootstrap-v7n52\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.407723 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.407792 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-credential-keys\") pod \"keystone-bootstrap-v7n52\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.407837 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-scripts\") pod \"keystone-bootstrap-v7n52\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.407872 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e22ff903-0c33-4be3-9905-d65afb2e6d22-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.407916 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-combined-ca-bundle\") pod \"keystone-bootstrap-v7n52\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.407942 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-config-data\") pod \"keystone-bootstrap-v7n52\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.407968 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e22ff903-0c33-4be3-9905-d65afb2e6d22-config-data\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.435525 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-credential-keys\") pod \"keystone-bootstrap-v7n52\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.435601 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-config-data\") pod \"keystone-bootstrap-v7n52\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.435950 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-fernet-keys\") pod \"keystone-bootstrap-v7n52\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.436628 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-scripts\") pod \"keystone-bootstrap-v7n52\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.445888 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-combined-ca-bundle\") pod \"keystone-bootstrap-v7n52\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.445996 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-dw8tc"] Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.448117 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9fd6\" (UniqueName: \"kubernetes.io/projected/77f95c7a-ce32-4353-8d7f-c5710d0b0057-kube-api-access-d9fd6\") pod \"keystone-bootstrap-v7n52\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.498395 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-8r767" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.515457 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.518848 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.518991 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e22ff903-0c33-4be3-9905-d65afb2e6d22-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.519050 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e22ff903-0c33-4be3-9905-d65afb2e6d22-config-data\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.519071 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfm7p\" (UniqueName: \"kubernetes.io/projected/e22ff903-0c33-4be3-9905-d65afb2e6d22-kube-api-access-lfm7p\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.519125 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e22ff903-0c33-4be3-9905-d65afb2e6d22-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.519145 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e22ff903-0c33-4be3-9905-d65afb2e6d22-logs\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.519162 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e22ff903-0c33-4be3-9905-d65afb2e6d22-scripts\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.521324 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e22ff903-0c33-4be3-9905-d65afb2e6d22-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.521674 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.522245 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e22ff903-0c33-4be3-9905-d65afb2e6d22-logs\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.524108 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.525046 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.525071 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9f50bff41a79b6548a20cce2c4162bc1d0a0a91c484e04e21b6d749439a0f41c/globalmount\"" pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.529800 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e22ff903-0c33-4be3-9905-d65afb2e6d22-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.538039 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e22ff903-0c33-4be3-9905-d65afb2e6d22-scripts\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.541913 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36677fa3-dfde-47b2-a5bb-524f18a3bc91" path="/var/lib/kubelet/pods/36677fa3-dfde-47b2-a5bb-524f18a3bc91/volumes" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.551117 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfm7p\" (UniqueName: \"kubernetes.io/projected/e22ff903-0c33-4be3-9905-d65afb2e6d22-kube-api-access-lfm7p\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.577271 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e22ff903-0c33-4be3-9905-d65afb2e6d22-config-data\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.615104 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\") pod \"glance-default-external-api-0\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.654201 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.656972 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.662233 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.662692 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-9x8k8" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.673690 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.716533 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.733458 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/169d952a-22c6-4299-ba37-73d7c6c4213d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.733711 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/169d952a-22c6-4299-ba37-73d7c6c4213d-logs\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.733787 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/169d952a-22c6-4299-ba37-73d7c6c4213d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.733966 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/169d952a-22c6-4299-ba37-73d7c6c4213d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.734097 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/169d952a-22c6-4299-ba37-73d7c6c4213d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.734174 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6z49s\" (UniqueName: \"kubernetes.io/projected/169d952a-22c6-4299-ba37-73d7c6c4213d-kube-api-access-6z49s\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.734262 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-44a42b08-9517-405f-9959-95693093d9c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.764165 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.824931 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-rrj89"] Dec 27 08:04:19 crc kubenswrapper[4934]: E1227 08:04:19.825822 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d97a13d-c9b8-4629-b847-0c578cdb71ff" containerName="init" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.825839 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d97a13d-c9b8-4629-b847-0c578cdb71ff" containerName="init" Dec 27 08:04:19 crc kubenswrapper[4934]: E1227 08:04:19.825849 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d97a13d-c9b8-4629-b847-0c578cdb71ff" containerName="dnsmasq-dns" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.825856 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d97a13d-c9b8-4629-b847-0c578cdb71ff" containerName="dnsmasq-dns" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.826356 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d97a13d-c9b8-4629-b847-0c578cdb71ff" containerName="dnsmasq-dns" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.830280 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rrj89" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.832148 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-zb2fh" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.837370 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-ovsdbserver-nb\") pod \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.837519 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-dns-swift-storage-0\") pod \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.837586 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drqh2\" (UniqueName: \"kubernetes.io/projected/3d97a13d-c9b8-4629-b847-0c578cdb71ff-kube-api-access-drqh2\") pod \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.837640 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-config\") pod \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.837707 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-ovsdbserver-sb\") pod \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.837724 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-dns-svc\") pod \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\" (UID: \"3d97a13d-c9b8-4629-b847-0c578cdb71ff\") " Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.838068 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/169d952a-22c6-4299-ba37-73d7c6c4213d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.838125 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/169d952a-22c6-4299-ba37-73d7c6c4213d-logs\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.838143 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/169d952a-22c6-4299-ba37-73d7c6c4213d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.838229 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/169d952a-22c6-4299-ba37-73d7c6c4213d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.838272 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/169d952a-22c6-4299-ba37-73d7c6c4213d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.838292 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6z49s\" (UniqueName: \"kubernetes.io/projected/169d952a-22c6-4299-ba37-73d7c6c4213d-kube-api-access-6z49s\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.838322 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-44a42b08-9517-405f-9959-95693093d9c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.840483 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.840644 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.846508 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/169d952a-22c6-4299-ba37-73d7c6c4213d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.853759 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/169d952a-22c6-4299-ba37-73d7c6c4213d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.854007 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/169d952a-22c6-4299-ba37-73d7c6c4213d-logs\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.860041 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/169d952a-22c6-4299-ba37-73d7c6c4213d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.869403 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-rrj89"] Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.870644 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d97a13d-c9b8-4629-b847-0c578cdb71ff-kube-api-access-drqh2" (OuterVolumeSpecName: "kube-api-access-drqh2") pod "3d97a13d-c9b8-4629-b847-0c578cdb71ff" (UID: "3d97a13d-c9b8-4629-b847-0c578cdb71ff"). InnerVolumeSpecName "kube-api-access-drqh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.886562 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.886604 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-44a42b08-9517-405f-9959-95693093d9c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/05b75074d2eec1bd1175d6431a19a8ca1381b55f1d6f90d62e5d62974d613123/globalmount\"" pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.907524 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/169d952a-22c6-4299-ba37-73d7c6c4213d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.922298 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6z49s\" (UniqueName: \"kubernetes.io/projected/169d952a-22c6-4299-ba37-73d7c6c4213d-kube-api-access-6z49s\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.945998 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-db-sync-config-data\") pod \"cinder-db-sync-rrj89\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " pod="openstack/cinder-db-sync-rrj89" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.946124 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8mxj\" (UniqueName: \"kubernetes.io/projected/72d0ed22-31be-4048-b031-139856c584c5-kube-api-access-l8mxj\") pod \"cinder-db-sync-rrj89\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " pod="openstack/cinder-db-sync-rrj89" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.946211 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/72d0ed22-31be-4048-b031-139856c584c5-etc-machine-id\") pod \"cinder-db-sync-rrj89\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " pod="openstack/cinder-db-sync-rrj89" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.946304 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-scripts\") pod \"cinder-db-sync-rrj89\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " pod="openstack/cinder-db-sync-rrj89" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.946346 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-combined-ca-bundle\") pod \"cinder-db-sync-rrj89\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " pod="openstack/cinder-db-sync-rrj89" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.946361 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-config-data\") pod \"cinder-db-sync-rrj89\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " pod="openstack/cinder-db-sync-rrj89" Dec 27 08:04:19 crc kubenswrapper[4934]: I1227 08:04:19.946449 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drqh2\" (UniqueName: \"kubernetes.io/projected/3d97a13d-c9b8-4629-b847-0c578cdb71ff-kube-api-access-drqh2\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.016450 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-44a42b08-9517-405f-9959-95693093d9c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0\") pod \"glance-default-internal-api-0\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.019761 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3d97a13d-c9b8-4629-b847-0c578cdb71ff" (UID: "3d97a13d-c9b8-4629-b847-0c578cdb71ff"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.049012 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/72d0ed22-31be-4048-b031-139856c584c5-etc-machine-id\") pod \"cinder-db-sync-rrj89\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " pod="openstack/cinder-db-sync-rrj89" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.049504 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-scripts\") pod \"cinder-db-sync-rrj89\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " pod="openstack/cinder-db-sync-rrj89" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.049552 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-combined-ca-bundle\") pod \"cinder-db-sync-rrj89\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " pod="openstack/cinder-db-sync-rrj89" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.049615 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-config-data\") pod \"cinder-db-sync-rrj89\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " pod="openstack/cinder-db-sync-rrj89" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.049699 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-db-sync-config-data\") pod \"cinder-db-sync-rrj89\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " pod="openstack/cinder-db-sync-rrj89" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.049771 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8mxj\" (UniqueName: \"kubernetes.io/projected/72d0ed22-31be-4048-b031-139856c584c5-kube-api-access-l8mxj\") pod \"cinder-db-sync-rrj89\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " pod="openstack/cinder-db-sync-rrj89" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.049879 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3d97a13d-c9b8-4629-b847-0c578cdb71ff" (UID: "3d97a13d-c9b8-4629-b847-0c578cdb71ff"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.049903 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.056819 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-config-data\") pod \"cinder-db-sync-rrj89\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " pod="openstack/cinder-db-sync-rrj89" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.056870 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/72d0ed22-31be-4048-b031-139856c584c5-etc-machine-id\") pod \"cinder-db-sync-rrj89\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " pod="openstack/cinder-db-sync-rrj89" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.057777 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-config" (OuterVolumeSpecName: "config") pod "3d97a13d-c9b8-4629-b847-0c578cdb71ff" (UID: "3d97a13d-c9b8-4629-b847-0c578cdb71ff"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.065693 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-db-sync-config-data\") pod \"cinder-db-sync-rrj89\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " pod="openstack/cinder-db-sync-rrj89" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.068823 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-combined-ca-bundle\") pod \"cinder-db-sync-rrj89\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " pod="openstack/cinder-db-sync-rrj89" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.086403 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3d97a13d-c9b8-4629-b847-0c578cdb71ff" (UID: "3d97a13d-c9b8-4629-b847-0c578cdb71ff"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.089065 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-scripts\") pod \"cinder-db-sync-rrj89\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " pod="openstack/cinder-db-sync-rrj89" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.092510 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.095017 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-f745s"] Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.096433 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-f745s" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.100617 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.100983 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-bx627" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.113270 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8mxj\" (UniqueName: \"kubernetes.io/projected/72d0ed22-31be-4048-b031-139856c584c5-kube-api-access-l8mxj\") pod \"cinder-db-sync-rrj89\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " pod="openstack/cinder-db-sync-rrj89" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.117152 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-f745s"] Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.171011 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfm8h\" (UniqueName: \"kubernetes.io/projected/5ef9a635-4d25-46d5-aa28-6c75f3cf8af4-kube-api-access-wfm8h\") pod \"barbican-db-sync-f745s\" (UID: \"5ef9a635-4d25-46d5-aa28-6c75f3cf8af4\") " pod="openstack/barbican-db-sync-f745s" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.171165 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ef9a635-4d25-46d5-aa28-6c75f3cf8af4-combined-ca-bundle\") pod \"barbican-db-sync-f745s\" (UID: \"5ef9a635-4d25-46d5-aa28-6c75f3cf8af4\") " pod="openstack/barbican-db-sync-f745s" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.171194 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5ef9a635-4d25-46d5-aa28-6c75f3cf8af4-db-sync-config-data\") pod \"barbican-db-sync-f745s\" (UID: \"5ef9a635-4d25-46d5-aa28-6c75f3cf8af4\") " pod="openstack/barbican-db-sync-f745s" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.171266 4934 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.171280 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.171289 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.191277 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3d97a13d-c9b8-4629-b847-0c578cdb71ff" (UID: "3d97a13d-c9b8-4629-b847-0c578cdb71ff"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.191838 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rrj89" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.210046 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-zpfs9"] Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.211660 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-zpfs9" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.215783 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.216130 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-88q97" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.215287 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.218156 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-xpk49" event={"ID":"3d97a13d-c9b8-4629-b847-0c578cdb71ff","Type":"ContainerDied","Data":"357fe674f6f3090b04c39c3223b0cb51d459cd002734c58d3bbbf3d6ad20e2d2"} Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.218199 4934 scope.go:117] "RemoveContainer" containerID="40bd45abc7cf4b892b27a1e4eec63f76876fbc3deba5ca067b98e910781c04ee" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.218309 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-xpk49" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.218971 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-zpfs9"] Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.238812 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" event={"ID":"1df64eee-76a8-4be2-8b36-665bc0f1564f","Type":"ContainerStarted","Data":"0a31e676527400f5082b959a90c929559a51b09287f384910eef01bfcc64300d"} Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.238853 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" event={"ID":"1df64eee-76a8-4be2-8b36-665bc0f1564f","Type":"ContainerStarted","Data":"895bd359c1243b6a62cfe2d688eb1f44e222bf71ad62c64a706630b35365899c"} Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.255534 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-v7n52"] Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.276873 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfm8h\" (UniqueName: \"kubernetes.io/projected/5ef9a635-4d25-46d5-aa28-6c75f3cf8af4-kube-api-access-wfm8h\") pod \"barbican-db-sync-f745s\" (UID: \"5ef9a635-4d25-46d5-aa28-6c75f3cf8af4\") " pod="openstack/barbican-db-sync-f745s" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.277245 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ef9a635-4d25-46d5-aa28-6c75f3cf8af4-combined-ca-bundle\") pod \"barbican-db-sync-f745s\" (UID: \"5ef9a635-4d25-46d5-aa28-6c75f3cf8af4\") " pod="openstack/barbican-db-sync-f745s" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.277294 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5ef9a635-4d25-46d5-aa28-6c75f3cf8af4-db-sync-config-data\") pod \"barbican-db-sync-f745s\" (UID: \"5ef9a635-4d25-46d5-aa28-6c75f3cf8af4\") " pod="openstack/barbican-db-sync-f745s" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.277437 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3d97a13d-c9b8-4629-b847-0c578cdb71ff-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.291649 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ef9a635-4d25-46d5-aa28-6c75f3cf8af4-combined-ca-bundle\") pod \"barbican-db-sync-f745s\" (UID: \"5ef9a635-4d25-46d5-aa28-6c75f3cf8af4\") " pod="openstack/barbican-db-sync-f745s" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.295652 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5ef9a635-4d25-46d5-aa28-6c75f3cf8af4-db-sync-config-data\") pod \"barbican-db-sync-f745s\" (UID: \"5ef9a635-4d25-46d5-aa28-6c75f3cf8af4\") " pod="openstack/barbican-db-sync-f745s" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.312853 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfm8h\" (UniqueName: \"kubernetes.io/projected/5ef9a635-4d25-46d5-aa28-6c75f3cf8af4-kube-api-access-wfm8h\") pod \"barbican-db-sync-f745s\" (UID: \"5ef9a635-4d25-46d5-aa28-6c75f3cf8af4\") " pod="openstack/barbican-db-sync-f745s" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.322289 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-xpk49"] Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.328043 4934 scope.go:117] "RemoveContainer" containerID="e4d8910c5b76c302c878aa5bac21827e5be05c7f33652ac0692b056f2f80d08a" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.335800 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-xpk49"] Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.382628 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38b7d77d-aea4-43e4-b5a9-4021d2562e4c-combined-ca-bundle\") pod \"neutron-db-sync-zpfs9\" (UID: \"38b7d77d-aea4-43e4-b5a9-4021d2562e4c\") " pod="openstack/neutron-db-sync-zpfs9" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.382910 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/38b7d77d-aea4-43e4-b5a9-4021d2562e4c-config\") pod \"neutron-db-sync-zpfs9\" (UID: \"38b7d77d-aea4-43e4-b5a9-4021d2562e4c\") " pod="openstack/neutron-db-sync-zpfs9" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.382992 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsbrh\" (UniqueName: \"kubernetes.io/projected/38b7d77d-aea4-43e4-b5a9-4021d2562e4c-kube-api-access-qsbrh\") pod \"neutron-db-sync-zpfs9\" (UID: \"38b7d77d-aea4-43e4-b5a9-4021d2562e4c\") " pod="openstack/neutron-db-sync-zpfs9" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.480570 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-f745s" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.485105 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38b7d77d-aea4-43e4-b5a9-4021d2562e4c-combined-ca-bundle\") pod \"neutron-db-sync-zpfs9\" (UID: \"38b7d77d-aea4-43e4-b5a9-4021d2562e4c\") " pod="openstack/neutron-db-sync-zpfs9" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.485843 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/38b7d77d-aea4-43e4-b5a9-4021d2562e4c-config\") pod \"neutron-db-sync-zpfs9\" (UID: \"38b7d77d-aea4-43e4-b5a9-4021d2562e4c\") " pod="openstack/neutron-db-sync-zpfs9" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.485919 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsbrh\" (UniqueName: \"kubernetes.io/projected/38b7d77d-aea4-43e4-b5a9-4021d2562e4c-kube-api-access-qsbrh\") pod \"neutron-db-sync-zpfs9\" (UID: \"38b7d77d-aea4-43e4-b5a9-4021d2562e4c\") " pod="openstack/neutron-db-sync-zpfs9" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.490266 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.497578 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38b7d77d-aea4-43e4-b5a9-4021d2562e4c-combined-ca-bundle\") pod \"neutron-db-sync-zpfs9\" (UID: \"38b7d77d-aea4-43e4-b5a9-4021d2562e4c\") " pod="openstack/neutron-db-sync-zpfs9" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.503043 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsbrh\" (UniqueName: \"kubernetes.io/projected/38b7d77d-aea4-43e4-b5a9-4021d2562e4c-kube-api-access-qsbrh\") pod \"neutron-db-sync-zpfs9\" (UID: \"38b7d77d-aea4-43e4-b5a9-4021d2562e4c\") " pod="openstack/neutron-db-sync-zpfs9" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.507511 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/38b7d77d-aea4-43e4-b5a9-4021d2562e4c-config\") pod \"neutron-db-sync-zpfs9\" (UID: \"38b7d77d-aea4-43e4-b5a9-4021d2562e4c\") " pod="openstack/neutron-db-sync-zpfs9" Dec 27 08:04:20 crc kubenswrapper[4934]: W1227 08:04:20.559693 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode22ff903_0c33_4be3_9905_d65afb2e6d22.slice/crio-c78fb4f7c5e77a0afe524ca9c526f030dccf91174801627330f48aa9152f34d1 WatchSource:0}: Error finding container c78fb4f7c5e77a0afe524ca9c526f030dccf91174801627330f48aa9152f34d1: Status 404 returned error can't find the container with id c78fb4f7c5e77a0afe524ca9c526f030dccf91174801627330f48aa9152f34d1 Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.644721 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-zpfs9" Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.841574 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 27 08:04:20 crc kubenswrapper[4934]: I1227 08:04:20.936643 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-rrj89"] Dec 27 08:04:21 crc kubenswrapper[4934]: I1227 08:04:21.160786 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-f745s"] Dec 27 08:04:21 crc kubenswrapper[4934]: I1227 08:04:21.344416 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v7n52" event={"ID":"77f95c7a-ce32-4353-8d7f-c5710d0b0057","Type":"ContainerStarted","Data":"7ee08fd9f1536658fe694fdf3478f7814115d9d18792f1ac758353bb2affd864"} Dec 27 08:04:21 crc kubenswrapper[4934]: I1227 08:04:21.344475 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v7n52" event={"ID":"77f95c7a-ce32-4353-8d7f-c5710d0b0057","Type":"ContainerStarted","Data":"a4f201e17f961cc2e0839ee2aeca718b23a297679ad86badd8a0bbe439c4cece"} Dec 27 08:04:21 crc kubenswrapper[4934]: I1227 08:04:21.413546 4934 generic.go:334] "Generic (PLEG): container finished" podID="1df64eee-76a8-4be2-8b36-665bc0f1564f" containerID="0a31e676527400f5082b959a90c929559a51b09287f384910eef01bfcc64300d" exitCode=0 Dec 27 08:04:21 crc kubenswrapper[4934]: I1227 08:04:21.413812 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" event={"ID":"1df64eee-76a8-4be2-8b36-665bc0f1564f","Type":"ContainerDied","Data":"0a31e676527400f5082b959a90c929559a51b09287f384910eef01bfcc64300d"} Dec 27 08:04:21 crc kubenswrapper[4934]: I1227 08:04:21.413838 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" event={"ID":"1df64eee-76a8-4be2-8b36-665bc0f1564f","Type":"ContainerStarted","Data":"d3369af297081db2f9235d0094a43a80f5646ff1b6ceca3e41271bb4dc40ab4f"} Dec 27 08:04:21 crc kubenswrapper[4934]: I1227 08:04:21.413877 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:21 crc kubenswrapper[4934]: I1227 08:04:21.414889 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-v7n52" podStartSLOduration=2.414865836 podStartE2EDuration="2.414865836s" podCreationTimestamp="2025-12-27 08:04:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:04:21.413701326 +0000 UTC m=+1322.234141920" watchObservedRunningTime="2025-12-27 08:04:21.414865836 +0000 UTC m=+1322.235306430" Dec 27 08:04:21 crc kubenswrapper[4934]: I1227 08:04:21.431391 4934 generic.go:334] "Generic (PLEG): container finished" podID="f7e0de58-4866-4ba7-9a80-7bce4eb52b97" containerID="c3f86223e2285fd2790d23cc1157f81d5862840766502576d5b20ca6aac68bd2" exitCode=0 Dec 27 08:04:21 crc kubenswrapper[4934]: I1227 08:04:21.431479 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-vkcbn" event={"ID":"f7e0de58-4866-4ba7-9a80-7bce4eb52b97","Type":"ContainerDied","Data":"c3f86223e2285fd2790d23cc1157f81d5862840766502576d5b20ca6aac68bd2"} Dec 27 08:04:21 crc kubenswrapper[4934]: I1227 08:04:21.445335 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e22ff903-0c33-4be3-9905-d65afb2e6d22","Type":"ContainerStarted","Data":"c78fb4f7c5e77a0afe524ca9c526f030dccf91174801627330f48aa9152f34d1"} Dec 27 08:04:21 crc kubenswrapper[4934]: I1227 08:04:21.458794 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" podStartSLOduration=3.458772627 podStartE2EDuration="3.458772627s" podCreationTimestamp="2025-12-27 08:04:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:04:21.443314769 +0000 UTC m=+1322.263755363" watchObservedRunningTime="2025-12-27 08:04:21.458772627 +0000 UTC m=+1322.279213221" Dec 27 08:04:21 crc kubenswrapper[4934]: I1227 08:04:21.501618 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d97a13d-c9b8-4629-b847-0c578cdb71ff" path="/var/lib/kubelet/pods/3d97a13d-c9b8-4629-b847-0c578cdb71ff/volumes" Dec 27 08:04:21 crc kubenswrapper[4934]: W1227 08:04:21.777184 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72d0ed22_31be_4048_b031_139856c584c5.slice/crio-4fc61ed7d15598ee35bfff4f23aede2f58447a5d5682e3e1cdc3d21129cc516b WatchSource:0}: Error finding container 4fc61ed7d15598ee35bfff4f23aede2f58447a5d5682e3e1cdc3d21129cc516b: Status 404 returned error can't find the container with id 4fc61ed7d15598ee35bfff4f23aede2f58447a5d5682e3e1cdc3d21129cc516b Dec 27 08:04:22 crc kubenswrapper[4934]: I1227 08:04:22.251014 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-zpfs9"] Dec 27 08:04:22 crc kubenswrapper[4934]: I1227 08:04:22.461897 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rrj89" event={"ID":"72d0ed22-31be-4048-b031-139856c584c5","Type":"ContainerStarted","Data":"4fc61ed7d15598ee35bfff4f23aede2f58447a5d5682e3e1cdc3d21129cc516b"} Dec 27 08:04:22 crc kubenswrapper[4934]: I1227 08:04:22.479024 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e22ff903-0c33-4be3-9905-d65afb2e6d22","Type":"ContainerStarted","Data":"8591239a223d2dffff925db7e1db2f211b983bc7f3cae26d182439464b2d3225"} Dec 27 08:04:22 crc kubenswrapper[4934]: I1227 08:04:22.499539 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-zpfs9" event={"ID":"38b7d77d-aea4-43e4-b5a9-4021d2562e4c","Type":"ContainerStarted","Data":"e4ca73aad4d3c626492002698fb4f4abfe8d862fe00362c088cb7f52b523eac4"} Dec 27 08:04:22 crc kubenswrapper[4934]: I1227 08:04:22.499585 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-zpfs9" event={"ID":"38b7d77d-aea4-43e4-b5a9-4021d2562e4c","Type":"ContainerStarted","Data":"b962efd15c45370beec85ceb72bccf6669f84cc53dd15d3267d968b837092654"} Dec 27 08:04:22 crc kubenswrapper[4934]: I1227 08:04:22.517045 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-f745s" event={"ID":"5ef9a635-4d25-46d5-aa28-6c75f3cf8af4","Type":"ContainerStarted","Data":"b65dc62674e3ade81c1c692855a0e1e48d48174edf673179ad47a36f0dba0e2e"} Dec 27 08:04:22 crc kubenswrapper[4934]: I1227 08:04:22.519325 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"169d952a-22c6-4299-ba37-73d7c6c4213d","Type":"ContainerStarted","Data":"8187c4267f80b09d02ddc178597b961030232509e24198d27a3d738324270034"} Dec 27 08:04:22 crc kubenswrapper[4934]: I1227 08:04:22.522746 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-zpfs9" podStartSLOduration=2.522722024 podStartE2EDuration="2.522722024s" podCreationTimestamp="2025-12-27 08:04:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:04:22.516583171 +0000 UTC m=+1323.337023765" watchObservedRunningTime="2025-12-27 08:04:22.522722024 +0000 UTC m=+1323.343162618" Dec 27 08:04:22 crc kubenswrapper[4934]: I1227 08:04:22.528054 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"badf697c-e657-4b80-9483-63c94c87afaf","Type":"ContainerStarted","Data":"63ebcede11cc284f2621953206a69b0ab47a59fa73cafc0cd32b945c5b87bc0f"} Dec 27 08:04:22 crc kubenswrapper[4934]: I1227 08:04:22.571702 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 27 08:04:22 crc kubenswrapper[4934]: I1227 08:04:22.651475 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.123265 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-vkcbn" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.228923 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-scripts\") pod \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\" (UID: \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\") " Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.229033 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-logs\") pod \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\" (UID: \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\") " Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.229104 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-config-data\") pod \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\" (UID: \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\") " Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.229142 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-combined-ca-bundle\") pod \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\" (UID: \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\") " Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.229234 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9wfl\" (UniqueName: \"kubernetes.io/projected/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-kube-api-access-s9wfl\") pod \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\" (UID: \"f7e0de58-4866-4ba7-9a80-7bce4eb52b97\") " Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.229859 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-logs" (OuterVolumeSpecName: "logs") pod "f7e0de58-4866-4ba7-9a80-7bce4eb52b97" (UID: "f7e0de58-4866-4ba7-9a80-7bce4eb52b97"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.235404 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-kube-api-access-s9wfl" (OuterVolumeSpecName: "kube-api-access-s9wfl") pod "f7e0de58-4866-4ba7-9a80-7bce4eb52b97" (UID: "f7e0de58-4866-4ba7-9a80-7bce4eb52b97"). InnerVolumeSpecName "kube-api-access-s9wfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.252734 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-scripts" (OuterVolumeSpecName: "scripts") pod "f7e0de58-4866-4ba7-9a80-7bce4eb52b97" (UID: "f7e0de58-4866-4ba7-9a80-7bce4eb52b97"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.265227 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-config-data" (OuterVolumeSpecName: "config-data") pod "f7e0de58-4866-4ba7-9a80-7bce4eb52b97" (UID: "f7e0de58-4866-4ba7-9a80-7bce4eb52b97"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.277253 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7e0de58-4866-4ba7-9a80-7bce4eb52b97" (UID: "f7e0de58-4866-4ba7-9a80-7bce4eb52b97"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.331248 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.331281 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-logs\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.331292 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.331301 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.331312 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9wfl\" (UniqueName: \"kubernetes.io/projected/f7e0de58-4866-4ba7-9a80-7bce4eb52b97-kube-api-access-s9wfl\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.565897 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e22ff903-0c33-4be3-9905-d65afb2e6d22","Type":"ContainerStarted","Data":"e5c4e2f9bead2b6e93fac04642ea846ed0e0cc7bf3a973eed38d994f7e9e2a5b"} Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.566117 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e22ff903-0c33-4be3-9905-d65afb2e6d22" containerName="glance-log" containerID="cri-o://8591239a223d2dffff925db7e1db2f211b983bc7f3cae26d182439464b2d3225" gracePeriod=30 Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.567230 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e22ff903-0c33-4be3-9905-d65afb2e6d22" containerName="glance-httpd" containerID="cri-o://e5c4e2f9bead2b6e93fac04642ea846ed0e0cc7bf3a973eed38d994f7e9e2a5b" gracePeriod=30 Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.594291 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5c6bd9586d-nf2kk"] Dec 27 08:04:23 crc kubenswrapper[4934]: E1227 08:04:23.595191 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7e0de58-4866-4ba7-9a80-7bce4eb52b97" containerName="placement-db-sync" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.595213 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7e0de58-4866-4ba7-9a80-7bce4eb52b97" containerName="placement-db-sync" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.596237 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7e0de58-4866-4ba7-9a80-7bce4eb52b97" containerName="placement-db-sync" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.610541 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"169d952a-22c6-4299-ba37-73d7c6c4213d","Type":"ContainerStarted","Data":"0da820bd2b6b596c6a20f35beef1cc1beed05ab7055633adbc9d15b968439aee"} Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.619250 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.626155 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.627017 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.656548 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5c6bd9586d-nf2kk"] Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.731507 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-vkcbn" event={"ID":"f7e0de58-4866-4ba7-9a80-7bce4eb52b97","Type":"ContainerDied","Data":"85231d5f4dd5e43d6234efe94f61a710ac0adb3f4cfd038cc5c24fc483c96be2"} Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.731518 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-vkcbn" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.743782 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.743759153 podStartE2EDuration="5.743759153s" podCreationTimestamp="2025-12-27 08:04:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:04:23.621837215 +0000 UTC m=+1324.442277819" watchObservedRunningTime="2025-12-27 08:04:23.743759153 +0000 UTC m=+1324.564199747" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.750743 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85231d5f4dd5e43d6234efe94f61a710ac0adb3f4cfd038cc5c24fc483c96be2" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.789161 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5259d7a2-5596-4cbf-88f1-2b0d59cf2b74-scripts\") pod \"placement-5c6bd9586d-nf2kk\" (UID: \"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74\") " pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.789407 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5259d7a2-5596-4cbf-88f1-2b0d59cf2b74-public-tls-certs\") pod \"placement-5c6bd9586d-nf2kk\" (UID: \"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74\") " pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.789508 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5259d7a2-5596-4cbf-88f1-2b0d59cf2b74-internal-tls-certs\") pod \"placement-5c6bd9586d-nf2kk\" (UID: \"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74\") " pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.789579 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf5tx\" (UniqueName: \"kubernetes.io/projected/5259d7a2-5596-4cbf-88f1-2b0d59cf2b74-kube-api-access-lf5tx\") pod \"placement-5c6bd9586d-nf2kk\" (UID: \"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74\") " pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.789608 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5259d7a2-5596-4cbf-88f1-2b0d59cf2b74-config-data\") pod \"placement-5c6bd9586d-nf2kk\" (UID: \"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74\") " pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.789627 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5259d7a2-5596-4cbf-88f1-2b0d59cf2b74-combined-ca-bundle\") pod \"placement-5c6bd9586d-nf2kk\" (UID: \"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74\") " pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.789714 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5259d7a2-5596-4cbf-88f1-2b0d59cf2b74-logs\") pod \"placement-5c6bd9586d-nf2kk\" (UID: \"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74\") " pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.892403 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5259d7a2-5596-4cbf-88f1-2b0d59cf2b74-internal-tls-certs\") pod \"placement-5c6bd9586d-nf2kk\" (UID: \"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74\") " pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.892486 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf5tx\" (UniqueName: \"kubernetes.io/projected/5259d7a2-5596-4cbf-88f1-2b0d59cf2b74-kube-api-access-lf5tx\") pod \"placement-5c6bd9586d-nf2kk\" (UID: \"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74\") " pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.892525 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5259d7a2-5596-4cbf-88f1-2b0d59cf2b74-config-data\") pod \"placement-5c6bd9586d-nf2kk\" (UID: \"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74\") " pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.892544 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5259d7a2-5596-4cbf-88f1-2b0d59cf2b74-combined-ca-bundle\") pod \"placement-5c6bd9586d-nf2kk\" (UID: \"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74\") " pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.892604 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5259d7a2-5596-4cbf-88f1-2b0d59cf2b74-logs\") pod \"placement-5c6bd9586d-nf2kk\" (UID: \"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74\") " pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.892654 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5259d7a2-5596-4cbf-88f1-2b0d59cf2b74-scripts\") pod \"placement-5c6bd9586d-nf2kk\" (UID: \"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74\") " pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.892758 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5259d7a2-5596-4cbf-88f1-2b0d59cf2b74-public-tls-certs\") pod \"placement-5c6bd9586d-nf2kk\" (UID: \"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74\") " pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.895868 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5259d7a2-5596-4cbf-88f1-2b0d59cf2b74-logs\") pod \"placement-5c6bd9586d-nf2kk\" (UID: \"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74\") " pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.919773 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5259d7a2-5596-4cbf-88f1-2b0d59cf2b74-config-data\") pod \"placement-5c6bd9586d-nf2kk\" (UID: \"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74\") " pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.925482 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5259d7a2-5596-4cbf-88f1-2b0d59cf2b74-scripts\") pod \"placement-5c6bd9586d-nf2kk\" (UID: \"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74\") " pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.926137 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5259d7a2-5596-4cbf-88f1-2b0d59cf2b74-public-tls-certs\") pod \"placement-5c6bd9586d-nf2kk\" (UID: \"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74\") " pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.958026 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5259d7a2-5596-4cbf-88f1-2b0d59cf2b74-internal-tls-certs\") pod \"placement-5c6bd9586d-nf2kk\" (UID: \"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74\") " pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.958584 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5259d7a2-5596-4cbf-88f1-2b0d59cf2b74-combined-ca-bundle\") pod \"placement-5c6bd9586d-nf2kk\" (UID: \"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74\") " pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.966489 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf5tx\" (UniqueName: \"kubernetes.io/projected/5259d7a2-5596-4cbf-88f1-2b0d59cf2b74-kube-api-access-lf5tx\") pod \"placement-5c6bd9586d-nf2kk\" (UID: \"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74\") " pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:23 crc kubenswrapper[4934]: I1227 08:04:23.971819 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.436580 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.508762 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e22ff903-0c33-4be3-9905-d65afb2e6d22-combined-ca-bundle\") pod \"e22ff903-0c33-4be3-9905-d65afb2e6d22\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.508829 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e22ff903-0c33-4be3-9905-d65afb2e6d22-logs\") pod \"e22ff903-0c33-4be3-9905-d65afb2e6d22\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.508929 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e22ff903-0c33-4be3-9905-d65afb2e6d22-scripts\") pod \"e22ff903-0c33-4be3-9905-d65afb2e6d22\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.508952 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e22ff903-0c33-4be3-9905-d65afb2e6d22-httpd-run\") pod \"e22ff903-0c33-4be3-9905-d65afb2e6d22\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.509016 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfm7p\" (UniqueName: \"kubernetes.io/projected/e22ff903-0c33-4be3-9905-d65afb2e6d22-kube-api-access-lfm7p\") pod \"e22ff903-0c33-4be3-9905-d65afb2e6d22\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.509163 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\") pod \"e22ff903-0c33-4be3-9905-d65afb2e6d22\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.509224 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e22ff903-0c33-4be3-9905-d65afb2e6d22-config-data\") pod \"e22ff903-0c33-4be3-9905-d65afb2e6d22\" (UID: \"e22ff903-0c33-4be3-9905-d65afb2e6d22\") " Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.509558 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e22ff903-0c33-4be3-9905-d65afb2e6d22-logs" (OuterVolumeSpecName: "logs") pod "e22ff903-0c33-4be3-9905-d65afb2e6d22" (UID: "e22ff903-0c33-4be3-9905-d65afb2e6d22"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.509989 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e22ff903-0c33-4be3-9905-d65afb2e6d22-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e22ff903-0c33-4be3-9905-d65afb2e6d22" (UID: "e22ff903-0c33-4be3-9905-d65afb2e6d22"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.510070 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e22ff903-0c33-4be3-9905-d65afb2e6d22-logs\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.516280 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e22ff903-0c33-4be3-9905-d65afb2e6d22-kube-api-access-lfm7p" (OuterVolumeSpecName: "kube-api-access-lfm7p") pod "e22ff903-0c33-4be3-9905-d65afb2e6d22" (UID: "e22ff903-0c33-4be3-9905-d65afb2e6d22"). InnerVolumeSpecName "kube-api-access-lfm7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.520005 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e22ff903-0c33-4be3-9905-d65afb2e6d22-scripts" (OuterVolumeSpecName: "scripts") pod "e22ff903-0c33-4be3-9905-d65afb2e6d22" (UID: "e22ff903-0c33-4be3-9905-d65afb2e6d22"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.547427 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6" (OuterVolumeSpecName: "glance") pod "e22ff903-0c33-4be3-9905-d65afb2e6d22" (UID: "e22ff903-0c33-4be3-9905-d65afb2e6d22"). InnerVolumeSpecName "pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.559617 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e22ff903-0c33-4be3-9905-d65afb2e6d22-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e22ff903-0c33-4be3-9905-d65afb2e6d22" (UID: "e22ff903-0c33-4be3-9905-d65afb2e6d22"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.599986 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e22ff903-0c33-4be3-9905-d65afb2e6d22-config-data" (OuterVolumeSpecName: "config-data") pod "e22ff903-0c33-4be3-9905-d65afb2e6d22" (UID: "e22ff903-0c33-4be3-9905-d65afb2e6d22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.614369 4934 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e22ff903-0c33-4be3-9905-d65afb2e6d22-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.615342 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfm7p\" (UniqueName: \"kubernetes.io/projected/e22ff903-0c33-4be3-9905-d65afb2e6d22-kube-api-access-lfm7p\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.615407 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\") on node \"crc\" " Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.615425 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e22ff903-0c33-4be3-9905-d65afb2e6d22-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.615438 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e22ff903-0c33-4be3-9905-d65afb2e6d22-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.615449 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e22ff903-0c33-4be3-9905-d65afb2e6d22-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.649136 4934 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.649281 4934 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6") on node "crc" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.701141 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5c6bd9586d-nf2kk"] Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.717221 4934 reconciler_common.go:293] "Volume detached for volume \"pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.743229 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.752128 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.758796 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5c6bd9586d-nf2kk" event={"ID":"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74","Type":"ContainerStarted","Data":"e62ad81b859350f2ad145fbf0a4a8867fabf6590d39520bd01a2e90934192aa5"} Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.763180 4934 generic.go:334] "Generic (PLEG): container finished" podID="e22ff903-0c33-4be3-9905-d65afb2e6d22" containerID="e5c4e2f9bead2b6e93fac04642ea846ed0e0cc7bf3a973eed38d994f7e9e2a5b" exitCode=143 Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.763207 4934 generic.go:334] "Generic (PLEG): container finished" podID="e22ff903-0c33-4be3-9905-d65afb2e6d22" containerID="8591239a223d2dffff925db7e1db2f211b983bc7f3cae26d182439464b2d3225" exitCode=143 Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.763239 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e22ff903-0c33-4be3-9905-d65afb2e6d22","Type":"ContainerDied","Data":"e5c4e2f9bead2b6e93fac04642ea846ed0e0cc7bf3a973eed38d994f7e9e2a5b"} Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.763266 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e22ff903-0c33-4be3-9905-d65afb2e6d22","Type":"ContainerDied","Data":"8591239a223d2dffff925db7e1db2f211b983bc7f3cae26d182439464b2d3225"} Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.763283 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e22ff903-0c33-4be3-9905-d65afb2e6d22","Type":"ContainerDied","Data":"c78fb4f7c5e77a0afe524ca9c526f030dccf91174801627330f48aa9152f34d1"} Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.763301 4934 scope.go:117] "RemoveContainer" containerID="e5c4e2f9bead2b6e93fac04642ea846ed0e0cc7bf3a973eed38d994f7e9e2a5b" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.763439 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.903701 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.921638 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.948008 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 27 08:04:24 crc kubenswrapper[4934]: E1227 08:04:24.948726 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e22ff903-0c33-4be3-9905-d65afb2e6d22" containerName="glance-httpd" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.948750 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="e22ff903-0c33-4be3-9905-d65afb2e6d22" containerName="glance-httpd" Dec 27 08:04:24 crc kubenswrapper[4934]: E1227 08:04:24.948812 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e22ff903-0c33-4be3-9905-d65afb2e6d22" containerName="glance-log" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.948821 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="e22ff903-0c33-4be3-9905-d65afb2e6d22" containerName="glance-log" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.949301 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="e22ff903-0c33-4be3-9905-d65afb2e6d22" containerName="glance-log" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.949358 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="e22ff903-0c33-4be3-9905-d65afb2e6d22" containerName="glance-httpd" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.952203 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.953174 4934 scope.go:117] "RemoveContainer" containerID="8591239a223d2dffff925db7e1db2f211b983bc7f3cae26d182439464b2d3225" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.954513 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 27 08:04:24 crc kubenswrapper[4934]: I1227 08:04:24.954795 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.000143 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.028638 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.028686 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f994d21a-55df-4ea2-a86d-41d18d10d4d1-logs\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.028727 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-config-data\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.028751 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-scripts\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.028779 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.028825 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfv77\" (UniqueName: \"kubernetes.io/projected/f994d21a-55df-4ea2-a86d-41d18d10d4d1-kube-api-access-xfv77\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.028908 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f994d21a-55df-4ea2-a86d-41d18d10d4d1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.028950 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.102683 4934 scope.go:117] "RemoveContainer" containerID="e5c4e2f9bead2b6e93fac04642ea846ed0e0cc7bf3a973eed38d994f7e9e2a5b" Dec 27 08:04:25 crc kubenswrapper[4934]: E1227 08:04:25.103536 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5c4e2f9bead2b6e93fac04642ea846ed0e0cc7bf3a973eed38d994f7e9e2a5b\": container with ID starting with e5c4e2f9bead2b6e93fac04642ea846ed0e0cc7bf3a973eed38d994f7e9e2a5b not found: ID does not exist" containerID="e5c4e2f9bead2b6e93fac04642ea846ed0e0cc7bf3a973eed38d994f7e9e2a5b" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.103576 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5c4e2f9bead2b6e93fac04642ea846ed0e0cc7bf3a973eed38d994f7e9e2a5b"} err="failed to get container status \"e5c4e2f9bead2b6e93fac04642ea846ed0e0cc7bf3a973eed38d994f7e9e2a5b\": rpc error: code = NotFound desc = could not find container \"e5c4e2f9bead2b6e93fac04642ea846ed0e0cc7bf3a973eed38d994f7e9e2a5b\": container with ID starting with e5c4e2f9bead2b6e93fac04642ea846ed0e0cc7bf3a973eed38d994f7e9e2a5b not found: ID does not exist" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.103600 4934 scope.go:117] "RemoveContainer" containerID="8591239a223d2dffff925db7e1db2f211b983bc7f3cae26d182439464b2d3225" Dec 27 08:04:25 crc kubenswrapper[4934]: E1227 08:04:25.104040 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8591239a223d2dffff925db7e1db2f211b983bc7f3cae26d182439464b2d3225\": container with ID starting with 8591239a223d2dffff925db7e1db2f211b983bc7f3cae26d182439464b2d3225 not found: ID does not exist" containerID="8591239a223d2dffff925db7e1db2f211b983bc7f3cae26d182439464b2d3225" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.104058 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8591239a223d2dffff925db7e1db2f211b983bc7f3cae26d182439464b2d3225"} err="failed to get container status \"8591239a223d2dffff925db7e1db2f211b983bc7f3cae26d182439464b2d3225\": rpc error: code = NotFound desc = could not find container \"8591239a223d2dffff925db7e1db2f211b983bc7f3cae26d182439464b2d3225\": container with ID starting with 8591239a223d2dffff925db7e1db2f211b983bc7f3cae26d182439464b2d3225 not found: ID does not exist" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.104070 4934 scope.go:117] "RemoveContainer" containerID="e5c4e2f9bead2b6e93fac04642ea846ed0e0cc7bf3a973eed38d994f7e9e2a5b" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.104398 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5c4e2f9bead2b6e93fac04642ea846ed0e0cc7bf3a973eed38d994f7e9e2a5b"} err="failed to get container status \"e5c4e2f9bead2b6e93fac04642ea846ed0e0cc7bf3a973eed38d994f7e9e2a5b\": rpc error: code = NotFound desc = could not find container \"e5c4e2f9bead2b6e93fac04642ea846ed0e0cc7bf3a973eed38d994f7e9e2a5b\": container with ID starting with e5c4e2f9bead2b6e93fac04642ea846ed0e0cc7bf3a973eed38d994f7e9e2a5b not found: ID does not exist" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.104418 4934 scope.go:117] "RemoveContainer" containerID="8591239a223d2dffff925db7e1db2f211b983bc7f3cae26d182439464b2d3225" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.104711 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8591239a223d2dffff925db7e1db2f211b983bc7f3cae26d182439464b2d3225"} err="failed to get container status \"8591239a223d2dffff925db7e1db2f211b983bc7f3cae26d182439464b2d3225\": rpc error: code = NotFound desc = could not find container \"8591239a223d2dffff925db7e1db2f211b983bc7f3cae26d182439464b2d3225\": container with ID starting with 8591239a223d2dffff925db7e1db2f211b983bc7f3cae26d182439464b2d3225 not found: ID does not exist" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.130768 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-config-data\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.130821 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-scripts\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.130854 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.130877 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfv77\" (UniqueName: \"kubernetes.io/projected/f994d21a-55df-4ea2-a86d-41d18d10d4d1-kube-api-access-xfv77\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.131036 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f994d21a-55df-4ea2-a86d-41d18d10d4d1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.131113 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.131299 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.131327 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f994d21a-55df-4ea2-a86d-41d18d10d4d1-logs\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.131797 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f994d21a-55df-4ea2-a86d-41d18d10d4d1-logs\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.136968 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f994d21a-55df-4ea2-a86d-41d18d10d4d1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.141497 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.141704 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.142987 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-config-data\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.143490 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-scripts\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.151203 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.151284 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9f50bff41a79b6548a20cce2c4162bc1d0a0a91c484e04e21b6d749439a0f41c/globalmount\"" pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.164208 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfv77\" (UniqueName: \"kubernetes.io/projected/f994d21a-55df-4ea2-a86d-41d18d10d4d1-kube-api-access-xfv77\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.239663 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\") pod \"glance-default-external-api-0\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.306361 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.494798 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e22ff903-0c33-4be3-9905-d65afb2e6d22" path="/var/lib/kubelet/pods/e22ff903-0c33-4be3-9905-d65afb2e6d22/volumes" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.805404 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"169d952a-22c6-4299-ba37-73d7c6c4213d","Type":"ContainerStarted","Data":"de63512a1adc2e0b64084fcc9acc0422e64f4629310539b620f3164206de5578"} Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.806412 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="169d952a-22c6-4299-ba37-73d7c6c4213d" containerName="glance-log" containerID="cri-o://0da820bd2b6b596c6a20f35beef1cc1beed05ab7055633adbc9d15b968439aee" gracePeriod=30 Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.806612 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="169d952a-22c6-4299-ba37-73d7c6c4213d" containerName="glance-httpd" containerID="cri-o://de63512a1adc2e0b64084fcc9acc0422e64f4629310539b620f3164206de5578" gracePeriod=30 Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.823939 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.838698 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.83867368 podStartE2EDuration="7.83867368s" podCreationTimestamp="2025-12-27 08:04:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:04:25.833979232 +0000 UTC m=+1326.654419836" watchObservedRunningTime="2025-12-27 08:04:25.83867368 +0000 UTC m=+1326.659114274" Dec 27 08:04:25 crc kubenswrapper[4934]: I1227 08:04:25.987757 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.589426 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.703354 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0\") pod \"169d952a-22c6-4299-ba37-73d7c6c4213d\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.703580 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/169d952a-22c6-4299-ba37-73d7c6c4213d-combined-ca-bundle\") pod \"169d952a-22c6-4299-ba37-73d7c6c4213d\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.703654 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6z49s\" (UniqueName: \"kubernetes.io/projected/169d952a-22c6-4299-ba37-73d7c6c4213d-kube-api-access-6z49s\") pod \"169d952a-22c6-4299-ba37-73d7c6c4213d\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.703737 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/169d952a-22c6-4299-ba37-73d7c6c4213d-logs\") pod \"169d952a-22c6-4299-ba37-73d7c6c4213d\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.703815 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/169d952a-22c6-4299-ba37-73d7c6c4213d-scripts\") pod \"169d952a-22c6-4299-ba37-73d7c6c4213d\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.703879 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/169d952a-22c6-4299-ba37-73d7c6c4213d-config-data\") pod \"169d952a-22c6-4299-ba37-73d7c6c4213d\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.703907 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/169d952a-22c6-4299-ba37-73d7c6c4213d-httpd-run\") pod \"169d952a-22c6-4299-ba37-73d7c6c4213d\" (UID: \"169d952a-22c6-4299-ba37-73d7c6c4213d\") " Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.704953 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/169d952a-22c6-4299-ba37-73d7c6c4213d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "169d952a-22c6-4299-ba37-73d7c6c4213d" (UID: "169d952a-22c6-4299-ba37-73d7c6c4213d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.705951 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/169d952a-22c6-4299-ba37-73d7c6c4213d-logs" (OuterVolumeSpecName: "logs") pod "169d952a-22c6-4299-ba37-73d7c6c4213d" (UID: "169d952a-22c6-4299-ba37-73d7c6c4213d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.717464 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/169d952a-22c6-4299-ba37-73d7c6c4213d-kube-api-access-6z49s" (OuterVolumeSpecName: "kube-api-access-6z49s") pod "169d952a-22c6-4299-ba37-73d7c6c4213d" (UID: "169d952a-22c6-4299-ba37-73d7c6c4213d"). InnerVolumeSpecName "kube-api-access-6z49s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.718489 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/169d952a-22c6-4299-ba37-73d7c6c4213d-scripts" (OuterVolumeSpecName: "scripts") pod "169d952a-22c6-4299-ba37-73d7c6c4213d" (UID: "169d952a-22c6-4299-ba37-73d7c6c4213d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.738519 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0" (OuterVolumeSpecName: "glance") pod "169d952a-22c6-4299-ba37-73d7c6c4213d" (UID: "169d952a-22c6-4299-ba37-73d7c6c4213d"). InnerVolumeSpecName "pvc-44a42b08-9517-405f-9959-95693093d9c0". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.769307 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/169d952a-22c6-4299-ba37-73d7c6c4213d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "169d952a-22c6-4299-ba37-73d7c6c4213d" (UID: "169d952a-22c6-4299-ba37-73d7c6c4213d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.806952 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/169d952a-22c6-4299-ba37-73d7c6c4213d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.806984 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6z49s\" (UniqueName: \"kubernetes.io/projected/169d952a-22c6-4299-ba37-73d7c6c4213d-kube-api-access-6z49s\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.807010 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/169d952a-22c6-4299-ba37-73d7c6c4213d-logs\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.807020 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/169d952a-22c6-4299-ba37-73d7c6c4213d-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.807028 4934 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/169d952a-22c6-4299-ba37-73d7c6c4213d-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.807050 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-44a42b08-9517-405f-9959-95693093d9c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0\") on node \"crc\" " Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.843993 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/169d952a-22c6-4299-ba37-73d7c6c4213d-config-data" (OuterVolumeSpecName: "config-data") pod "169d952a-22c6-4299-ba37-73d7c6c4213d" (UID: "169d952a-22c6-4299-ba37-73d7c6c4213d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.875111 4934 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.875780 4934 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-44a42b08-9517-405f-9959-95693093d9c0" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0") on node "crc" Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.911073 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/169d952a-22c6-4299-ba37-73d7c6c4213d-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.911114 4934 reconciler_common.go:293] "Volume detached for volume \"pvc-44a42b08-9517-405f-9959-95693093d9c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.923709 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5c6bd9586d-nf2kk" event={"ID":"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74","Type":"ContainerStarted","Data":"12a051ed486e7d8b758717f2b27abd88808d9f87122f5b2626f3fbd80cfb3d08"} Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.924209 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5c6bd9586d-nf2kk" event={"ID":"5259d7a2-5596-4cbf-88f1-2b0d59cf2b74","Type":"ContainerStarted","Data":"d78c49fec46321a26a01b6172a91aaceac88df8aa814f9a7bf1537dc86b3600c"} Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.924232 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.925609 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.933248 4934 generic.go:334] "Generic (PLEG): container finished" podID="169d952a-22c6-4299-ba37-73d7c6c4213d" containerID="de63512a1adc2e0b64084fcc9acc0422e64f4629310539b620f3164206de5578" exitCode=143 Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.933275 4934 generic.go:334] "Generic (PLEG): container finished" podID="169d952a-22c6-4299-ba37-73d7c6c4213d" containerID="0da820bd2b6b596c6a20f35beef1cc1beed05ab7055633adbc9d15b968439aee" exitCode=143 Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.933334 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"169d952a-22c6-4299-ba37-73d7c6c4213d","Type":"ContainerDied","Data":"de63512a1adc2e0b64084fcc9acc0422e64f4629310539b620f3164206de5578"} Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.933358 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"169d952a-22c6-4299-ba37-73d7c6c4213d","Type":"ContainerDied","Data":"0da820bd2b6b596c6a20f35beef1cc1beed05ab7055633adbc9d15b968439aee"} Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.933371 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"169d952a-22c6-4299-ba37-73d7c6c4213d","Type":"ContainerDied","Data":"8187c4267f80b09d02ddc178597b961030232509e24198d27a3d738324270034"} Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.933386 4934 scope.go:117] "RemoveContainer" containerID="de63512a1adc2e0b64084fcc9acc0422e64f4629310539b620f3164206de5578" Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.933541 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.953461 4934 generic.go:334] "Generic (PLEG): container finished" podID="77f95c7a-ce32-4353-8d7f-c5710d0b0057" containerID="7ee08fd9f1536658fe694fdf3478f7814115d9d18792f1ac758353bb2affd864" exitCode=0 Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.953618 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v7n52" event={"ID":"77f95c7a-ce32-4353-8d7f-c5710d0b0057","Type":"ContainerDied","Data":"7ee08fd9f1536658fe694fdf3478f7814115d9d18792f1ac758353bb2affd864"} Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.959455 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f994d21a-55df-4ea2-a86d-41d18d10d4d1","Type":"ContainerStarted","Data":"5aa76dc8fd33bf617d82ad0e757ea43d3b9145049183265764b584931927502a"} Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.966130 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5c6bd9586d-nf2kk" podStartSLOduration=3.96610787 podStartE2EDuration="3.96610787s" podCreationTimestamp="2025-12-27 08:04:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:04:26.945340769 +0000 UTC m=+1327.765781363" watchObservedRunningTime="2025-12-27 08:04:26.96610787 +0000 UTC m=+1327.786548464" Dec 27 08:04:26 crc kubenswrapper[4934]: I1227 08:04:26.988751 4934 scope.go:117] "RemoveContainer" containerID="0da820bd2b6b596c6a20f35beef1cc1beed05ab7055633adbc9d15b968439aee" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.002318 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.013554 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.031130 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 27 08:04:27 crc kubenswrapper[4934]: E1227 08:04:27.031662 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="169d952a-22c6-4299-ba37-73d7c6c4213d" containerName="glance-httpd" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.031677 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="169d952a-22c6-4299-ba37-73d7c6c4213d" containerName="glance-httpd" Dec 27 08:04:27 crc kubenswrapper[4934]: E1227 08:04:27.031696 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="169d952a-22c6-4299-ba37-73d7c6c4213d" containerName="glance-log" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.031702 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="169d952a-22c6-4299-ba37-73d7c6c4213d" containerName="glance-log" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.031893 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="169d952a-22c6-4299-ba37-73d7c6c4213d" containerName="glance-log" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.031913 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="169d952a-22c6-4299-ba37-73d7c6c4213d" containerName="glance-httpd" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.033865 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.041903 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.058286 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.058483 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.101128 4934 scope.go:117] "RemoveContainer" containerID="de63512a1adc2e0b64084fcc9acc0422e64f4629310539b620f3164206de5578" Dec 27 08:04:27 crc kubenswrapper[4934]: E1227 08:04:27.116393 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de63512a1adc2e0b64084fcc9acc0422e64f4629310539b620f3164206de5578\": container with ID starting with de63512a1adc2e0b64084fcc9acc0422e64f4629310539b620f3164206de5578 not found: ID does not exist" containerID="de63512a1adc2e0b64084fcc9acc0422e64f4629310539b620f3164206de5578" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.116455 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de63512a1adc2e0b64084fcc9acc0422e64f4629310539b620f3164206de5578"} err="failed to get container status \"de63512a1adc2e0b64084fcc9acc0422e64f4629310539b620f3164206de5578\": rpc error: code = NotFound desc = could not find container \"de63512a1adc2e0b64084fcc9acc0422e64f4629310539b620f3164206de5578\": container with ID starting with de63512a1adc2e0b64084fcc9acc0422e64f4629310539b620f3164206de5578 not found: ID does not exist" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.116489 4934 scope.go:117] "RemoveContainer" containerID="0da820bd2b6b596c6a20f35beef1cc1beed05ab7055633adbc9d15b968439aee" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.118211 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-44a42b08-9517-405f-9959-95693093d9c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.118330 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86f3ebe1-9843-44e6-907a-e453aab4b9e8-logs\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.118367 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86f3ebe1-9843-44e6-907a-e453aab4b9e8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.118431 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmw7x\" (UniqueName: \"kubernetes.io/projected/86f3ebe1-9843-44e6-907a-e453aab4b9e8-kube-api-access-nmw7x\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.118497 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.118535 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.118592 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.118789 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: E1227 08:04:27.119096 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0da820bd2b6b596c6a20f35beef1cc1beed05ab7055633adbc9d15b968439aee\": container with ID starting with 0da820bd2b6b596c6a20f35beef1cc1beed05ab7055633adbc9d15b968439aee not found: ID does not exist" containerID="0da820bd2b6b596c6a20f35beef1cc1beed05ab7055633adbc9d15b968439aee" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.119135 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0da820bd2b6b596c6a20f35beef1cc1beed05ab7055633adbc9d15b968439aee"} err="failed to get container status \"0da820bd2b6b596c6a20f35beef1cc1beed05ab7055633adbc9d15b968439aee\": rpc error: code = NotFound desc = could not find container \"0da820bd2b6b596c6a20f35beef1cc1beed05ab7055633adbc9d15b968439aee\": container with ID starting with 0da820bd2b6b596c6a20f35beef1cc1beed05ab7055633adbc9d15b968439aee not found: ID does not exist" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.119162 4934 scope.go:117] "RemoveContainer" containerID="de63512a1adc2e0b64084fcc9acc0422e64f4629310539b620f3164206de5578" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.129986 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de63512a1adc2e0b64084fcc9acc0422e64f4629310539b620f3164206de5578"} err="failed to get container status \"de63512a1adc2e0b64084fcc9acc0422e64f4629310539b620f3164206de5578\": rpc error: code = NotFound desc = could not find container \"de63512a1adc2e0b64084fcc9acc0422e64f4629310539b620f3164206de5578\": container with ID starting with de63512a1adc2e0b64084fcc9acc0422e64f4629310539b620f3164206de5578 not found: ID does not exist" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.130027 4934 scope.go:117] "RemoveContainer" containerID="0da820bd2b6b596c6a20f35beef1cc1beed05ab7055633adbc9d15b968439aee" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.134765 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0da820bd2b6b596c6a20f35beef1cc1beed05ab7055633adbc9d15b968439aee"} err="failed to get container status \"0da820bd2b6b596c6a20f35beef1cc1beed05ab7055633adbc9d15b968439aee\": rpc error: code = NotFound desc = could not find container \"0da820bd2b6b596c6a20f35beef1cc1beed05ab7055633adbc9d15b968439aee\": container with ID starting with 0da820bd2b6b596c6a20f35beef1cc1beed05ab7055633adbc9d15b968439aee not found: ID does not exist" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.223540 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.224404 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-44a42b08-9517-405f-9959-95693093d9c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.225250 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86f3ebe1-9843-44e6-907a-e453aab4b9e8-logs\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.225280 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86f3ebe1-9843-44e6-907a-e453aab4b9e8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.225324 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmw7x\" (UniqueName: \"kubernetes.io/projected/86f3ebe1-9843-44e6-907a-e453aab4b9e8-kube-api-access-nmw7x\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.225365 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.225386 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.225407 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.226404 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86f3ebe1-9843-44e6-907a-e453aab4b9e8-logs\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.226569 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86f3ebe1-9843-44e6-907a-e453aab4b9e8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.229816 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.229925 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-44a42b08-9517-405f-9959-95693093d9c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/05b75074d2eec1bd1175d6431a19a8ca1381b55f1d6f90d62e5d62974d613123/globalmount\"" pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.230994 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.234020 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.234674 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.237125 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: E1227 08:04:27.258933 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod169d952a_22c6_4299_ba37_73d7c6c4213d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod169d952a_22c6_4299_ba37_73d7c6c4213d.slice/crio-8187c4267f80b09d02ddc178597b961030232509e24198d27a3d738324270034\": RecentStats: unable to find data in memory cache]" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.259328 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmw7x\" (UniqueName: \"kubernetes.io/projected/86f3ebe1-9843-44e6-907a-e453aab4b9e8-kube-api-access-nmw7x\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.290493 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-44a42b08-9517-405f-9959-95693093d9c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0\") pod \"glance-default-internal-api-0\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.405166 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.487354 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="169d952a-22c6-4299-ba37-73d7c6c4213d" path="/var/lib/kubelet/pods/169d952a-22c6-4299-ba37-73d7c6c4213d/volumes" Dec 27 08:04:27 crc kubenswrapper[4934]: I1227 08:04:27.982309 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f994d21a-55df-4ea2-a86d-41d18d10d4d1","Type":"ContainerStarted","Data":"5738140d2bba76b7bc6a908b15245ca158c60b158eba7d68b51d42af6c3a07b2"} Dec 27 08:04:28 crc kubenswrapper[4934]: I1227 08:04:28.808467 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:04:28 crc kubenswrapper[4934]: I1227 08:04:28.880173 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-5n5c8"] Dec 27 08:04:28 crc kubenswrapper[4934]: I1227 08:04:28.880381 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" podUID="9ba80d2d-2587-4da1-aa21-31fc5e2775b0" containerName="dnsmasq-dns" containerID="cri-o://6f020a4ec55892c99858e76f37cbc694ceac6da4d533f1f1836c0b62bee489f2" gracePeriod=10 Dec 27 08:04:29 crc kubenswrapper[4934]: I1227 08:04:29.006240 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f994d21a-55df-4ea2-a86d-41d18d10d4d1","Type":"ContainerStarted","Data":"d16fe7c3850179dc183ac942ae64cb6065521e25fa98f3bc7089f111a9b8c0a1"} Dec 27 08:04:29 crc kubenswrapper[4934]: I1227 08:04:29.029350 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.029334732 podStartE2EDuration="5.029334732s" podCreationTimestamp="2025-12-27 08:04:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:04:29.025575057 +0000 UTC m=+1329.846015651" watchObservedRunningTime="2025-12-27 08:04:29.029334732 +0000 UTC m=+1329.849775326" Dec 27 08:04:29 crc kubenswrapper[4934]: I1227 08:04:29.205756 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" podUID="9ba80d2d-2587-4da1-aa21-31fc5e2775b0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.150:5353: connect: connection refused" Dec 27 08:04:30 crc kubenswrapper[4934]: I1227 08:04:30.023183 4934 generic.go:334] "Generic (PLEG): container finished" podID="9ba80d2d-2587-4da1-aa21-31fc5e2775b0" containerID="6f020a4ec55892c99858e76f37cbc694ceac6da4d533f1f1836c0b62bee489f2" exitCode=0 Dec 27 08:04:30 crc kubenswrapper[4934]: I1227 08:04:30.023273 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" event={"ID":"9ba80d2d-2587-4da1-aa21-31fc5e2775b0","Type":"ContainerDied","Data":"6f020a4ec55892c99858e76f37cbc694ceac6da4d533f1f1836c0b62bee489f2"} Dec 27 08:04:34 crc kubenswrapper[4934]: I1227 08:04:34.206299 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" podUID="9ba80d2d-2587-4da1-aa21-31fc5e2775b0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.150:5353: connect: connection refused" Dec 27 08:04:34 crc kubenswrapper[4934]: I1227 08:04:34.556365 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:34 crc kubenswrapper[4934]: I1227 08:04:34.633017 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-combined-ca-bundle\") pod \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " Dec 27 08:04:34 crc kubenswrapper[4934]: I1227 08:04:34.633068 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-config-data\") pod \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " Dec 27 08:04:34 crc kubenswrapper[4934]: I1227 08:04:34.633179 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-credential-keys\") pod \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " Dec 27 08:04:34 crc kubenswrapper[4934]: I1227 08:04:34.633289 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9fd6\" (UniqueName: \"kubernetes.io/projected/77f95c7a-ce32-4353-8d7f-c5710d0b0057-kube-api-access-d9fd6\") pod \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " Dec 27 08:04:34 crc kubenswrapper[4934]: I1227 08:04:34.633401 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-fernet-keys\") pod \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " Dec 27 08:04:34 crc kubenswrapper[4934]: I1227 08:04:34.633499 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-scripts\") pod \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\" (UID: \"77f95c7a-ce32-4353-8d7f-c5710d0b0057\") " Dec 27 08:04:34 crc kubenswrapper[4934]: I1227 08:04:34.642200 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-scripts" (OuterVolumeSpecName: "scripts") pod "77f95c7a-ce32-4353-8d7f-c5710d0b0057" (UID: "77f95c7a-ce32-4353-8d7f-c5710d0b0057"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:34 crc kubenswrapper[4934]: I1227 08:04:34.686595 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77f95c7a-ce32-4353-8d7f-c5710d0b0057-kube-api-access-d9fd6" (OuterVolumeSpecName: "kube-api-access-d9fd6") pod "77f95c7a-ce32-4353-8d7f-c5710d0b0057" (UID: "77f95c7a-ce32-4353-8d7f-c5710d0b0057"). InnerVolumeSpecName "kube-api-access-d9fd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:04:34 crc kubenswrapper[4934]: I1227 08:04:34.688870 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "77f95c7a-ce32-4353-8d7f-c5710d0b0057" (UID: "77f95c7a-ce32-4353-8d7f-c5710d0b0057"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:34 crc kubenswrapper[4934]: I1227 08:04:34.688965 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "77f95c7a-ce32-4353-8d7f-c5710d0b0057" (UID: "77f95c7a-ce32-4353-8d7f-c5710d0b0057"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:34 crc kubenswrapper[4934]: I1227 08:04:34.697241 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77f95c7a-ce32-4353-8d7f-c5710d0b0057" (UID: "77f95c7a-ce32-4353-8d7f-c5710d0b0057"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:34 crc kubenswrapper[4934]: I1227 08:04:34.730916 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-config-data" (OuterVolumeSpecName: "config-data") pod "77f95c7a-ce32-4353-8d7f-c5710d0b0057" (UID: "77f95c7a-ce32-4353-8d7f-c5710d0b0057"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:34 crc kubenswrapper[4934]: I1227 08:04:34.736853 4934 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:34 crc kubenswrapper[4934]: I1227 08:04:34.736883 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:34 crc kubenswrapper[4934]: I1227 08:04:34.736894 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:34 crc kubenswrapper[4934]: I1227 08:04:34.736908 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:34 crc kubenswrapper[4934]: I1227 08:04:34.736916 4934 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/77f95c7a-ce32-4353-8d7f-c5710d0b0057-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:34 crc kubenswrapper[4934]: I1227 08:04:34.736926 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9fd6\" (UniqueName: \"kubernetes.io/projected/77f95c7a-ce32-4353-8d7f-c5710d0b0057-kube-api-access-d9fd6\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.098884 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-v7n52" event={"ID":"77f95c7a-ce32-4353-8d7f-c5710d0b0057","Type":"ContainerDied","Data":"a4f201e17f961cc2e0839ee2aeca718b23a297679ad86badd8a0bbe439c4cece"} Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.098932 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4f201e17f961cc2e0839ee2aeca718b23a297679ad86badd8a0bbe439c4cece" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.099450 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-v7n52" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.308542 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.308596 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.346471 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.358912 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.665719 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7fd8c76fd8-kbl58"] Dec 27 08:04:35 crc kubenswrapper[4934]: E1227 08:04:35.668422 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77f95c7a-ce32-4353-8d7f-c5710d0b0057" containerName="keystone-bootstrap" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.668652 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="77f95c7a-ce32-4353-8d7f-c5710d0b0057" containerName="keystone-bootstrap" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.669082 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="77f95c7a-ce32-4353-8d7f-c5710d0b0057" containerName="keystone-bootstrap" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.670142 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.672342 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.672915 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.672971 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-8r767" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.673038 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.673202 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.676907 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7fd8c76fd8-kbl58"] Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.678327 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.759262 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-combined-ca-bundle\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.759374 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-public-tls-certs\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.759429 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-internal-tls-certs\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.759457 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-credential-keys\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.759475 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h58cr\" (UniqueName: \"kubernetes.io/projected/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-kube-api-access-h58cr\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.759506 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-scripts\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.759543 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-config-data\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.759578 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-fernet-keys\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.861368 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-combined-ca-bundle\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.861451 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-public-tls-certs\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.861488 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-internal-tls-certs\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.861512 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-credential-keys\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.861532 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h58cr\" (UniqueName: \"kubernetes.io/projected/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-kube-api-access-h58cr\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.861566 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-scripts\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.861604 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-config-data\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.861635 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-fernet-keys\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.865686 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-fernet-keys\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.866453 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-credential-keys\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.866734 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-scripts\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.869653 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-config-data\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.869909 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-public-tls-certs\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.872648 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-combined-ca-bundle\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.879423 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-internal-tls-certs\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.879993 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h58cr\" (UniqueName: \"kubernetes.io/projected/1de4b3bd-106e-44ce-b68e-8a2c28916b9b-kube-api-access-h58cr\") pod \"keystone-7fd8c76fd8-kbl58\" (UID: \"1de4b3bd-106e-44ce-b68e-8a2c28916b9b\") " pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:35 crc kubenswrapper[4934]: I1227 08:04:35.994203 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:36 crc kubenswrapper[4934]: I1227 08:04:36.110037 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 27 08:04:36 crc kubenswrapper[4934]: I1227 08:04:36.110114 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 27 08:04:38 crc kubenswrapper[4934]: I1227 08:04:38.995097 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 27 08:04:38 crc kubenswrapper[4934]: I1227 08:04:38.995518 4934 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 27 08:04:39 crc kubenswrapper[4934]: I1227 08:04:39.055960 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 27 08:04:44 crc kubenswrapper[4934]: I1227 08:04:44.206358 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" podUID="9ba80d2d-2587-4da1-aa21-31fc5e2775b0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.150:5353: i/o timeout" Dec 27 08:04:44 crc kubenswrapper[4934]: I1227 08:04:44.207056 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:04:45 crc kubenswrapper[4934]: I1227 08:04:45.330568 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:04:45 crc kubenswrapper[4934]: I1227 08:04:45.330652 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:04:47 crc kubenswrapper[4934]: I1227 08:04:47.804749 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:04:47 crc kubenswrapper[4934]: I1227 08:04:47.966828 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9lhz\" (UniqueName: \"kubernetes.io/projected/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-kube-api-access-z9lhz\") pod \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\" (UID: \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\") " Dec 27 08:04:47 crc kubenswrapper[4934]: I1227 08:04:47.966965 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-dns-svc\") pod \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\" (UID: \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\") " Dec 27 08:04:47 crc kubenswrapper[4934]: I1227 08:04:47.967033 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-ovsdbserver-sb\") pod \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\" (UID: \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\") " Dec 27 08:04:47 crc kubenswrapper[4934]: I1227 08:04:47.967205 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-ovsdbserver-nb\") pod \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\" (UID: \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\") " Dec 27 08:04:47 crc kubenswrapper[4934]: I1227 08:04:47.967339 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-config\") pod \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\" (UID: \"9ba80d2d-2587-4da1-aa21-31fc5e2775b0\") " Dec 27 08:04:47 crc kubenswrapper[4934]: I1227 08:04:47.976278 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-kube-api-access-z9lhz" (OuterVolumeSpecName: "kube-api-access-z9lhz") pod "9ba80d2d-2587-4da1-aa21-31fc5e2775b0" (UID: "9ba80d2d-2587-4da1-aa21-31fc5e2775b0"). InnerVolumeSpecName "kube-api-access-z9lhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:04:48 crc kubenswrapper[4934]: I1227 08:04:48.020227 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9ba80d2d-2587-4da1-aa21-31fc5e2775b0" (UID: "9ba80d2d-2587-4da1-aa21-31fc5e2775b0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:04:48 crc kubenswrapper[4934]: I1227 08:04:48.029992 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9ba80d2d-2587-4da1-aa21-31fc5e2775b0" (UID: "9ba80d2d-2587-4da1-aa21-31fc5e2775b0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:04:48 crc kubenswrapper[4934]: I1227 08:04:48.044169 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9ba80d2d-2587-4da1-aa21-31fc5e2775b0" (UID: "9ba80d2d-2587-4da1-aa21-31fc5e2775b0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:04:48 crc kubenswrapper[4934]: I1227 08:04:48.049824 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-config" (OuterVolumeSpecName: "config") pod "9ba80d2d-2587-4da1-aa21-31fc5e2775b0" (UID: "9ba80d2d-2587-4da1-aa21-31fc5e2775b0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:04:48 crc kubenswrapper[4934]: I1227 08:04:48.070053 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:48 crc kubenswrapper[4934]: I1227 08:04:48.070130 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:48 crc kubenswrapper[4934]: I1227 08:04:48.070143 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9lhz\" (UniqueName: \"kubernetes.io/projected/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-kube-api-access-z9lhz\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:48 crc kubenswrapper[4934]: I1227 08:04:48.070153 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:48 crc kubenswrapper[4934]: I1227 08:04:48.070161 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ba80d2d-2587-4da1-aa21-31fc5e2775b0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:48 crc kubenswrapper[4934]: I1227 08:04:48.268408 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" event={"ID":"9ba80d2d-2587-4da1-aa21-31fc5e2775b0","Type":"ContainerDied","Data":"af2309b0d8f07ac64a5d58a21e2725160bc2f158f8735417342fd085b404641e"} Dec 27 08:04:48 crc kubenswrapper[4934]: I1227 08:04:48.268501 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" Dec 27 08:04:48 crc kubenswrapper[4934]: I1227 08:04:48.268861 4934 scope.go:117] "RemoveContainer" containerID="6f020a4ec55892c99858e76f37cbc694ceac6da4d533f1f1836c0b62bee489f2" Dec 27 08:04:48 crc kubenswrapper[4934]: I1227 08:04:48.316109 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-5n5c8"] Dec 27 08:04:48 crc kubenswrapper[4934]: I1227 08:04:48.327803 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-5n5c8"] Dec 27 08:04:48 crc kubenswrapper[4934]: E1227 08:04:48.383761 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 27 08:04:48 crc kubenswrapper[4934]: E1227 08:04:48.383956 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wfm8h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-f745s_openstack(5ef9a635-4d25-46d5-aa28-6c75f3cf8af4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:04:48 crc kubenswrapper[4934]: E1227 08:04:48.385795 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-f745s" podUID="5ef9a635-4d25-46d5-aa28-6c75f3cf8af4" Dec 27 08:04:49 crc kubenswrapper[4934]: I1227 08:04:49.207298 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-5n5c8" podUID="9ba80d2d-2587-4da1-aa21-31fc5e2775b0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.150:5353: i/o timeout" Dec 27 08:04:49 crc kubenswrapper[4934]: E1227 08:04:49.284324 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-f745s" podUID="5ef9a635-4d25-46d5-aa28-6c75f3cf8af4" Dec 27 08:04:49 crc kubenswrapper[4934]: I1227 08:04:49.485651 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ba80d2d-2587-4da1-aa21-31fc5e2775b0" path="/var/lib/kubelet/pods/9ba80d2d-2587-4da1-aa21-31fc5e2775b0/volumes" Dec 27 08:04:51 crc kubenswrapper[4934]: E1227 08:04:51.311340 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/sg-core:latest" Dec 27 08:04:51 crc kubenswrapper[4934]: E1227 08:04:51.311995 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:sg-core,Image:quay.io/openstack-k8s-operators/sg-core:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:sg-core-conf-yaml,ReadOnly:false,MountPath:/etc/sg-core.conf.yaml,SubPath:sg-core.conf.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vksgf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(badf697c-e657-4b80-9483-63c94c87afaf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:04:52 crc kubenswrapper[4934]: I1227 08:04:52.537117 4934 scope.go:117] "RemoveContainer" containerID="dc5c4a8cb477b83fb27aae7ca0babf37748a40f4d0ab4db1c056699a80f9bd1c" Dec 27 08:04:52 crc kubenswrapper[4934]: E1227 08:04:52.563894 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 27 08:04:52 crc kubenswrapper[4934]: E1227 08:04:52.564129 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l8mxj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-rrj89_openstack(72d0ed22-31be-4048-b031-139856c584c5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:04:52 crc kubenswrapper[4934]: E1227 08:04:52.565334 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-rrj89" podUID="72d0ed22-31be-4048-b031-139856c584c5" Dec 27 08:04:52 crc kubenswrapper[4934]: I1227 08:04:52.988458 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7fd8c76fd8-kbl58"] Dec 27 08:04:53 crc kubenswrapper[4934]: I1227 08:04:53.079573 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 27 08:04:53 crc kubenswrapper[4934]: I1227 08:04:53.340743 4934 generic.go:334] "Generic (PLEG): container finished" podID="38b7d77d-aea4-43e4-b5a9-4021d2562e4c" containerID="e4ca73aad4d3c626492002698fb4f4abfe8d862fe00362c088cb7f52b523eac4" exitCode=0 Dec 27 08:04:53 crc kubenswrapper[4934]: I1227 08:04:53.340984 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-zpfs9" event={"ID":"38b7d77d-aea4-43e4-b5a9-4021d2562e4c","Type":"ContainerDied","Data":"e4ca73aad4d3c626492002698fb4f4abfe8d862fe00362c088cb7f52b523eac4"} Dec 27 08:04:53 crc kubenswrapper[4934]: I1227 08:04:53.347570 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-5qllw" event={"ID":"61f5132a-303d-4f76-ae58-c57d82a59fd4","Type":"ContainerStarted","Data":"db4ed2f28d2e711d823b1ea76d20bc93ee4ecdf738544c01eaea7df7995d7b4a"} Dec 27 08:04:53 crc kubenswrapper[4934]: I1227 08:04:53.355452 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7fd8c76fd8-kbl58" event={"ID":"1de4b3bd-106e-44ce-b68e-8a2c28916b9b","Type":"ContainerStarted","Data":"7962d80388eca20850800af179884418cb3464bb5f481c543d0a29dc27a59a33"} Dec 27 08:04:53 crc kubenswrapper[4934]: I1227 08:04:53.355495 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7fd8c76fd8-kbl58" event={"ID":"1de4b3bd-106e-44ce-b68e-8a2c28916b9b","Type":"ContainerStarted","Data":"8215e6e3cee2baf5fa0407416202cdcaa16471e37b3440a44b494a35144e15b3"} Dec 27 08:04:53 crc kubenswrapper[4934]: I1227 08:04:53.357328 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:04:53 crc kubenswrapper[4934]: I1227 08:04:53.359143 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"86f3ebe1-9843-44e6-907a-e453aab4b9e8","Type":"ContainerStarted","Data":"5dfe586abfcd1ecfe60596eafc9089b4922ece508d866d8b6e575effe0cbbe18"} Dec 27 08:04:53 crc kubenswrapper[4934]: E1227 08:04:53.360957 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-rrj89" podUID="72d0ed22-31be-4048-b031-139856c584c5" Dec 27 08:04:53 crc kubenswrapper[4934]: I1227 08:04:53.383524 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7fd8c76fd8-kbl58" podStartSLOduration=18.383499582 podStartE2EDuration="18.383499582s" podCreationTimestamp="2025-12-27 08:04:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:04:53.379364238 +0000 UTC m=+1354.199804852" watchObservedRunningTime="2025-12-27 08:04:53.383499582 +0000 UTC m=+1354.203940176" Dec 27 08:04:53 crc kubenswrapper[4934]: I1227 08:04:53.408987 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-5qllw" podStartSLOduration=4.799706888 podStartE2EDuration="39.408964601s" podCreationTimestamp="2025-12-27 08:04:14 +0000 UTC" firstStartedPulling="2025-12-27 08:04:17.916224908 +0000 UTC m=+1318.736665502" lastFinishedPulling="2025-12-27 08:04:52.525482621 +0000 UTC m=+1353.345923215" observedRunningTime="2025-12-27 08:04:53.40177647 +0000 UTC m=+1354.222217074" watchObservedRunningTime="2025-12-27 08:04:53.408964601 +0000 UTC m=+1354.229405205" Dec 27 08:04:54 crc kubenswrapper[4934]: I1227 08:04:54.382354 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"86f3ebe1-9843-44e6-907a-e453aab4b9e8","Type":"ContainerStarted","Data":"2fd6b17ae0e178466b758988239a7b5ed70321c3d412fecab516a68a5df0d602"} Dec 27 08:04:54 crc kubenswrapper[4934]: I1227 08:04:54.786387 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-zpfs9" Dec 27 08:04:54 crc kubenswrapper[4934]: I1227 08:04:54.940364 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/38b7d77d-aea4-43e4-b5a9-4021d2562e4c-config\") pod \"38b7d77d-aea4-43e4-b5a9-4021d2562e4c\" (UID: \"38b7d77d-aea4-43e4-b5a9-4021d2562e4c\") " Dec 27 08:04:54 crc kubenswrapper[4934]: I1227 08:04:54.940598 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38b7d77d-aea4-43e4-b5a9-4021d2562e4c-combined-ca-bundle\") pod \"38b7d77d-aea4-43e4-b5a9-4021d2562e4c\" (UID: \"38b7d77d-aea4-43e4-b5a9-4021d2562e4c\") " Dec 27 08:04:54 crc kubenswrapper[4934]: I1227 08:04:54.941107 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsbrh\" (UniqueName: \"kubernetes.io/projected/38b7d77d-aea4-43e4-b5a9-4021d2562e4c-kube-api-access-qsbrh\") pod \"38b7d77d-aea4-43e4-b5a9-4021d2562e4c\" (UID: \"38b7d77d-aea4-43e4-b5a9-4021d2562e4c\") " Dec 27 08:04:54 crc kubenswrapper[4934]: I1227 08:04:54.946409 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38b7d77d-aea4-43e4-b5a9-4021d2562e4c-kube-api-access-qsbrh" (OuterVolumeSpecName: "kube-api-access-qsbrh") pod "38b7d77d-aea4-43e4-b5a9-4021d2562e4c" (UID: "38b7d77d-aea4-43e4-b5a9-4021d2562e4c"). InnerVolumeSpecName "kube-api-access-qsbrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:04:54 crc kubenswrapper[4934]: I1227 08:04:54.967355 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38b7d77d-aea4-43e4-b5a9-4021d2562e4c-config" (OuterVolumeSpecName: "config") pod "38b7d77d-aea4-43e4-b5a9-4021d2562e4c" (UID: "38b7d77d-aea4-43e4-b5a9-4021d2562e4c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:54 crc kubenswrapper[4934]: I1227 08:04:54.976008 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38b7d77d-aea4-43e4-b5a9-4021d2562e4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "38b7d77d-aea4-43e4-b5a9-4021d2562e4c" (UID: "38b7d77d-aea4-43e4-b5a9-4021d2562e4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.044092 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/38b7d77d-aea4-43e4-b5a9-4021d2562e4c-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.044125 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38b7d77d-aea4-43e4-b5a9-4021d2562e4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.044135 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsbrh\" (UniqueName: \"kubernetes.io/projected/38b7d77d-aea4-43e4-b5a9-4021d2562e4c-kube-api-access-qsbrh\") on node \"crc\" DevicePath \"\"" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.405315 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"86f3ebe1-9843-44e6-907a-e453aab4b9e8","Type":"ContainerStarted","Data":"9ee83e3a8e87b6f9a43e8c2d49fec8346411b9357abba2d50e8154b7814f4de2"} Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.411192 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-zpfs9" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.412986 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-zpfs9" event={"ID":"38b7d77d-aea4-43e4-b5a9-4021d2562e4c","Type":"ContainerDied","Data":"b962efd15c45370beec85ceb72bccf6669f84cc53dd15d3267d968b837092654"} Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.413022 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b962efd15c45370beec85ceb72bccf6669f84cc53dd15d3267d968b837092654" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.442560 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=29.442516509 podStartE2EDuration="29.442516509s" podCreationTimestamp="2025-12-27 08:04:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:04:55.431115863 +0000 UTC m=+1356.251556457" watchObservedRunningTime="2025-12-27 08:04:55.442516509 +0000 UTC m=+1356.262957103" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.850591 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-76g9s"] Dec 27 08:04:55 crc kubenswrapper[4934]: E1227 08:04:55.851061 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38b7d77d-aea4-43e4-b5a9-4021d2562e4c" containerName="neutron-db-sync" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.851076 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="38b7d77d-aea4-43e4-b5a9-4021d2562e4c" containerName="neutron-db-sync" Dec 27 08:04:55 crc kubenswrapper[4934]: E1227 08:04:55.851131 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ba80d2d-2587-4da1-aa21-31fc5e2775b0" containerName="dnsmasq-dns" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.851138 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ba80d2d-2587-4da1-aa21-31fc5e2775b0" containerName="dnsmasq-dns" Dec 27 08:04:55 crc kubenswrapper[4934]: E1227 08:04:55.851154 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ba80d2d-2587-4da1-aa21-31fc5e2775b0" containerName="init" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.851161 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ba80d2d-2587-4da1-aa21-31fc5e2775b0" containerName="init" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.851367 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="38b7d77d-aea4-43e4-b5a9-4021d2562e4c" containerName="neutron-db-sync" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.851387 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ba80d2d-2587-4da1-aa21-31fc5e2775b0" containerName="dnsmasq-dns" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.852485 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.868282 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-76g9s"] Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.935377 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-76886c86d8-c6mdr"] Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.937628 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76886c86d8-c6mdr" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.946466 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.946751 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.946829 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.946941 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-88q97" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.953301 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-76886c86d8-c6mdr"] Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.964328 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-dns-svc\") pod \"dnsmasq-dns-6b7b667979-76g9s\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.964380 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m62pd\" (UniqueName: \"kubernetes.io/projected/ea0027c6-81c4-4d65-96a6-43eb035d2043-kube-api-access-m62pd\") pod \"dnsmasq-dns-6b7b667979-76g9s\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.964431 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-76g9s\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.964491 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-76g9s\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.964508 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-config\") pod \"dnsmasq-dns-6b7b667979-76g9s\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:04:55 crc kubenswrapper[4934]: I1227 08:04:55.964583 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-76g9s\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.066171 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-combined-ca-bundle\") pod \"neutron-76886c86d8-c6mdr\" (UID: \"c67ffca3-2539-4344-848e-e251e564f42b\") " pod="openstack/neutron-76886c86d8-c6mdr" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.066228 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-config\") pod \"dnsmasq-dns-6b7b667979-76g9s\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.066252 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-76g9s\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.066522 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p5mf\" (UniqueName: \"kubernetes.io/projected/c67ffca3-2539-4344-848e-e251e564f42b-kube-api-access-8p5mf\") pod \"neutron-76886c86d8-c6mdr\" (UID: \"c67ffca3-2539-4344-848e-e251e564f42b\") " pod="openstack/neutron-76886c86d8-c6mdr" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.066595 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-76g9s\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.066709 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-httpd-config\") pod \"neutron-76886c86d8-c6mdr\" (UID: \"c67ffca3-2539-4344-848e-e251e564f42b\") " pod="openstack/neutron-76886c86d8-c6mdr" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.066871 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-dns-svc\") pod \"dnsmasq-dns-6b7b667979-76g9s\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.066999 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m62pd\" (UniqueName: \"kubernetes.io/projected/ea0027c6-81c4-4d65-96a6-43eb035d2043-kube-api-access-m62pd\") pod \"dnsmasq-dns-6b7b667979-76g9s\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.067157 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-ovndb-tls-certs\") pod \"neutron-76886c86d8-c6mdr\" (UID: \"c67ffca3-2539-4344-848e-e251e564f42b\") " pod="openstack/neutron-76886c86d8-c6mdr" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.067195 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-76g9s\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.067309 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-config\") pod \"neutron-76886c86d8-c6mdr\" (UID: \"c67ffca3-2539-4344-848e-e251e564f42b\") " pod="openstack/neutron-76886c86d8-c6mdr" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.067432 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-config\") pod \"dnsmasq-dns-6b7b667979-76g9s\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.067472 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-76g9s\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.067496 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-76g9s\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.067737 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-dns-svc\") pod \"dnsmasq-dns-6b7b667979-76g9s\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.067963 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-76g9s\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.086885 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m62pd\" (UniqueName: \"kubernetes.io/projected/ea0027c6-81c4-4d65-96a6-43eb035d2043-kube-api-access-m62pd\") pod \"dnsmasq-dns-6b7b667979-76g9s\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.168755 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.169403 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-ovndb-tls-certs\") pod \"neutron-76886c86d8-c6mdr\" (UID: \"c67ffca3-2539-4344-848e-e251e564f42b\") " pod="openstack/neutron-76886c86d8-c6mdr" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.169478 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-config\") pod \"neutron-76886c86d8-c6mdr\" (UID: \"c67ffca3-2539-4344-848e-e251e564f42b\") " pod="openstack/neutron-76886c86d8-c6mdr" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.169514 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-combined-ca-bundle\") pod \"neutron-76886c86d8-c6mdr\" (UID: \"c67ffca3-2539-4344-848e-e251e564f42b\") " pod="openstack/neutron-76886c86d8-c6mdr" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.169597 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p5mf\" (UniqueName: \"kubernetes.io/projected/c67ffca3-2539-4344-848e-e251e564f42b-kube-api-access-8p5mf\") pod \"neutron-76886c86d8-c6mdr\" (UID: \"c67ffca3-2539-4344-848e-e251e564f42b\") " pod="openstack/neutron-76886c86d8-c6mdr" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.169630 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-httpd-config\") pod \"neutron-76886c86d8-c6mdr\" (UID: \"c67ffca3-2539-4344-848e-e251e564f42b\") " pod="openstack/neutron-76886c86d8-c6mdr" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.178477 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-ovndb-tls-certs\") pod \"neutron-76886c86d8-c6mdr\" (UID: \"c67ffca3-2539-4344-848e-e251e564f42b\") " pod="openstack/neutron-76886c86d8-c6mdr" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.178744 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-config\") pod \"neutron-76886c86d8-c6mdr\" (UID: \"c67ffca3-2539-4344-848e-e251e564f42b\") " pod="openstack/neutron-76886c86d8-c6mdr" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.181780 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-combined-ca-bundle\") pod \"neutron-76886c86d8-c6mdr\" (UID: \"c67ffca3-2539-4344-848e-e251e564f42b\") " pod="openstack/neutron-76886c86d8-c6mdr" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.187474 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-httpd-config\") pod \"neutron-76886c86d8-c6mdr\" (UID: \"c67ffca3-2539-4344-848e-e251e564f42b\") " pod="openstack/neutron-76886c86d8-c6mdr" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.187742 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p5mf\" (UniqueName: \"kubernetes.io/projected/c67ffca3-2539-4344-848e-e251e564f42b-kube-api-access-8p5mf\") pod \"neutron-76886c86d8-c6mdr\" (UID: \"c67ffca3-2539-4344-848e-e251e564f42b\") " pod="openstack/neutron-76886c86d8-c6mdr" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.273624 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76886c86d8-c6mdr" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.294150 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.338195 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5c6bd9586d-nf2kk" Dec 27 08:04:56 crc kubenswrapper[4934]: I1227 08:04:56.724295 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-76g9s"] Dec 27 08:04:57 crc kubenswrapper[4934]: I1227 08:04:57.405755 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 27 08:04:57 crc kubenswrapper[4934]: I1227 08:04:57.405800 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 27 08:04:57 crc kubenswrapper[4934]: I1227 08:04:57.405809 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 27 08:04:57 crc kubenswrapper[4934]: I1227 08:04:57.405818 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 27 08:04:57 crc kubenswrapper[4934]: I1227 08:04:57.454509 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 27 08:04:57 crc kubenswrapper[4934]: I1227 08:04:57.459289 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 27 08:04:57 crc kubenswrapper[4934]: I1227 08:04:57.929162 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c689587cc-6glpp"] Dec 27 08:04:57 crc kubenswrapper[4934]: I1227 08:04:57.931114 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:57 crc kubenswrapper[4934]: I1227 08:04:57.933320 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 27 08:04:57 crc kubenswrapper[4934]: I1227 08:04:57.935114 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 27 08:04:57 crc kubenswrapper[4934]: I1227 08:04:57.943504 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c689587cc-6glpp"] Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.016563 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/72923f32-62d5-49f6-81dc-9741f36139e8-httpd-config\") pod \"neutron-c689587cc-6glpp\" (UID: \"72923f32-62d5-49f6-81dc-9741f36139e8\") " pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.016625 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/72923f32-62d5-49f6-81dc-9741f36139e8-config\") pod \"neutron-c689587cc-6glpp\" (UID: \"72923f32-62d5-49f6-81dc-9741f36139e8\") " pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.016716 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72923f32-62d5-49f6-81dc-9741f36139e8-public-tls-certs\") pod \"neutron-c689587cc-6glpp\" (UID: \"72923f32-62d5-49f6-81dc-9741f36139e8\") " pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.016773 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/72923f32-62d5-49f6-81dc-9741f36139e8-ovndb-tls-certs\") pod \"neutron-c689587cc-6glpp\" (UID: \"72923f32-62d5-49f6-81dc-9741f36139e8\") " pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.016799 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72923f32-62d5-49f6-81dc-9741f36139e8-combined-ca-bundle\") pod \"neutron-c689587cc-6glpp\" (UID: \"72923f32-62d5-49f6-81dc-9741f36139e8\") " pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.016839 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72923f32-62d5-49f6-81dc-9741f36139e8-internal-tls-certs\") pod \"neutron-c689587cc-6glpp\" (UID: \"72923f32-62d5-49f6-81dc-9741f36139e8\") " pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.016864 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jpdj\" (UniqueName: \"kubernetes.io/projected/72923f32-62d5-49f6-81dc-9741f36139e8-kube-api-access-9jpdj\") pod \"neutron-c689587cc-6glpp\" (UID: \"72923f32-62d5-49f6-81dc-9741f36139e8\") " pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.121659 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/72923f32-62d5-49f6-81dc-9741f36139e8-httpd-config\") pod \"neutron-c689587cc-6glpp\" (UID: \"72923f32-62d5-49f6-81dc-9741f36139e8\") " pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.121906 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/72923f32-62d5-49f6-81dc-9741f36139e8-config\") pod \"neutron-c689587cc-6glpp\" (UID: \"72923f32-62d5-49f6-81dc-9741f36139e8\") " pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.121993 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72923f32-62d5-49f6-81dc-9741f36139e8-public-tls-certs\") pod \"neutron-c689587cc-6glpp\" (UID: \"72923f32-62d5-49f6-81dc-9741f36139e8\") " pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.122052 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/72923f32-62d5-49f6-81dc-9741f36139e8-ovndb-tls-certs\") pod \"neutron-c689587cc-6glpp\" (UID: \"72923f32-62d5-49f6-81dc-9741f36139e8\") " pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.122101 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72923f32-62d5-49f6-81dc-9741f36139e8-combined-ca-bundle\") pod \"neutron-c689587cc-6glpp\" (UID: \"72923f32-62d5-49f6-81dc-9741f36139e8\") " pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.122140 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72923f32-62d5-49f6-81dc-9741f36139e8-internal-tls-certs\") pod \"neutron-c689587cc-6glpp\" (UID: \"72923f32-62d5-49f6-81dc-9741f36139e8\") " pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.122160 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jpdj\" (UniqueName: \"kubernetes.io/projected/72923f32-62d5-49f6-81dc-9741f36139e8-kube-api-access-9jpdj\") pod \"neutron-c689587cc-6glpp\" (UID: \"72923f32-62d5-49f6-81dc-9741f36139e8\") " pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.167169 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/72923f32-62d5-49f6-81dc-9741f36139e8-config\") pod \"neutron-c689587cc-6glpp\" (UID: \"72923f32-62d5-49f6-81dc-9741f36139e8\") " pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.169070 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jpdj\" (UniqueName: \"kubernetes.io/projected/72923f32-62d5-49f6-81dc-9741f36139e8-kube-api-access-9jpdj\") pod \"neutron-c689587cc-6glpp\" (UID: \"72923f32-62d5-49f6-81dc-9741f36139e8\") " pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.169519 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/72923f32-62d5-49f6-81dc-9741f36139e8-ovndb-tls-certs\") pod \"neutron-c689587cc-6glpp\" (UID: \"72923f32-62d5-49f6-81dc-9741f36139e8\") " pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.180766 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/72923f32-62d5-49f6-81dc-9741f36139e8-internal-tls-certs\") pod \"neutron-c689587cc-6glpp\" (UID: \"72923f32-62d5-49f6-81dc-9741f36139e8\") " pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.188062 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/72923f32-62d5-49f6-81dc-9741f36139e8-httpd-config\") pod \"neutron-c689587cc-6glpp\" (UID: \"72923f32-62d5-49f6-81dc-9741f36139e8\") " pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.197614 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/72923f32-62d5-49f6-81dc-9741f36139e8-public-tls-certs\") pod \"neutron-c689587cc-6glpp\" (UID: \"72923f32-62d5-49f6-81dc-9741f36139e8\") " pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.198921 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72923f32-62d5-49f6-81dc-9741f36139e8-combined-ca-bundle\") pod \"neutron-c689587cc-6glpp\" (UID: \"72923f32-62d5-49f6-81dc-9741f36139e8\") " pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:04:58 crc kubenswrapper[4934]: I1227 08:04:58.287221 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:05:00 crc kubenswrapper[4934]: I1227 08:05:00.506415 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-76g9s" event={"ID":"ea0027c6-81c4-4d65-96a6-43eb035d2043","Type":"ContainerStarted","Data":"2897983f242ebc3009030fc753cbef7e0671608abac9db84c943b84ea43d670b"} Dec 27 08:05:00 crc kubenswrapper[4934]: I1227 08:05:00.800478 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 27 08:05:02 crc kubenswrapper[4934]: I1227 08:05:02.971277 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c689587cc-6glpp"] Dec 27 08:05:03 crc kubenswrapper[4934]: I1227 08:05:03.155770 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 27 08:05:03 crc kubenswrapper[4934]: E1227 08:05:03.311273 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"sg-core\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="badf697c-e657-4b80-9483-63c94c87afaf" Dec 27 08:05:03 crc kubenswrapper[4934]: I1227 08:05:03.501935 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-76886c86d8-c6mdr"] Dec 27 08:05:03 crc kubenswrapper[4934]: I1227 08:05:03.543867 4934 generic.go:334] "Generic (PLEG): container finished" podID="ea0027c6-81c4-4d65-96a6-43eb035d2043" containerID="3dd317627a8b1981085d0b2a2bd00372273d1ad0d839ba67ada22bbbb49ef0f0" exitCode=0 Dec 27 08:05:03 crc kubenswrapper[4934]: I1227 08:05:03.544123 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-76g9s" event={"ID":"ea0027c6-81c4-4d65-96a6-43eb035d2043","Type":"ContainerDied","Data":"3dd317627a8b1981085d0b2a2bd00372273d1ad0d839ba67ada22bbbb49ef0f0"} Dec 27 08:05:03 crc kubenswrapper[4934]: I1227 08:05:03.552500 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"badf697c-e657-4b80-9483-63c94c87afaf","Type":"ContainerStarted","Data":"533777f7714ac651df3ce7fadf06b65fc0d7096fe9f0cb914d969c67314f0142"} Dec 27 08:05:03 crc kubenswrapper[4934]: I1227 08:05:03.552633 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="badf697c-e657-4b80-9483-63c94c87afaf" containerName="ceilometer-central-agent" containerID="cri-o://45a559573157263c959004566fc36f27391dce93ff6361826ad2d9438ebd9a66" gracePeriod=30 Dec 27 08:05:03 crc kubenswrapper[4934]: I1227 08:05:03.552821 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 27 08:05:03 crc kubenswrapper[4934]: I1227 08:05:03.552858 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="badf697c-e657-4b80-9483-63c94c87afaf" containerName="proxy-httpd" containerID="cri-o://533777f7714ac651df3ce7fadf06b65fc0d7096fe9f0cb914d969c67314f0142" gracePeriod=30 Dec 27 08:05:03 crc kubenswrapper[4934]: I1227 08:05:03.552898 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="badf697c-e657-4b80-9483-63c94c87afaf" containerName="ceilometer-notification-agent" containerID="cri-o://63ebcede11cc284f2621953206a69b0ab47a59fa73cafc0cd32b945c5b87bc0f" gracePeriod=30 Dec 27 08:05:03 crc kubenswrapper[4934]: I1227 08:05:03.562385 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76886c86d8-c6mdr" event={"ID":"c67ffca3-2539-4344-848e-e251e564f42b","Type":"ContainerStarted","Data":"3fc3110b2306014fade5fc8825927ac431b1dbf9b36adf451772a256afe222c4"} Dec 27 08:05:03 crc kubenswrapper[4934]: I1227 08:05:03.570053 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c689587cc-6glpp" event={"ID":"72923f32-62d5-49f6-81dc-9741f36139e8","Type":"ContainerStarted","Data":"00aa35d6a2d167dce09d30b378d8ffc69496baf966b28f325d4dc699be085f1e"} Dec 27 08:05:03 crc kubenswrapper[4934]: I1227 08:05:03.570112 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c689587cc-6glpp" event={"ID":"72923f32-62d5-49f6-81dc-9741f36139e8","Type":"ContainerStarted","Data":"1559e5fe15787a3a39ab5d85d7a6022e24e208eefe66e8e62bf43634bafc8b37"} Dec 27 08:05:04 crc kubenswrapper[4934]: I1227 08:05:04.593424 4934 generic.go:334] "Generic (PLEG): container finished" podID="badf697c-e657-4b80-9483-63c94c87afaf" containerID="533777f7714ac651df3ce7fadf06b65fc0d7096fe9f0cb914d969c67314f0142" exitCode=0 Dec 27 08:05:04 crc kubenswrapper[4934]: I1227 08:05:04.593878 4934 generic.go:334] "Generic (PLEG): container finished" podID="badf697c-e657-4b80-9483-63c94c87afaf" containerID="45a559573157263c959004566fc36f27391dce93ff6361826ad2d9438ebd9a66" exitCode=0 Dec 27 08:05:04 crc kubenswrapper[4934]: I1227 08:05:04.593487 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"badf697c-e657-4b80-9483-63c94c87afaf","Type":"ContainerDied","Data":"533777f7714ac651df3ce7fadf06b65fc0d7096fe9f0cb914d969c67314f0142"} Dec 27 08:05:04 crc kubenswrapper[4934]: I1227 08:05:04.593964 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"badf697c-e657-4b80-9483-63c94c87afaf","Type":"ContainerDied","Data":"45a559573157263c959004566fc36f27391dce93ff6361826ad2d9438ebd9a66"} Dec 27 08:05:04 crc kubenswrapper[4934]: I1227 08:05:04.596807 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76886c86d8-c6mdr" event={"ID":"c67ffca3-2539-4344-848e-e251e564f42b","Type":"ContainerStarted","Data":"67e0ab891a2e8f56faeae88e28a6fca9233b0b966fcab3583810609696d05004"} Dec 27 08:05:04 crc kubenswrapper[4934]: I1227 08:05:04.596852 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76886c86d8-c6mdr" event={"ID":"c67ffca3-2539-4344-848e-e251e564f42b","Type":"ContainerStarted","Data":"267c584dc50d95618fa637ec47ff3c93f9e2ea0d158207d0591b2838c4afa90d"} Dec 27 08:05:04 crc kubenswrapper[4934]: I1227 08:05:04.598409 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-76886c86d8-c6mdr" Dec 27 08:05:04 crc kubenswrapper[4934]: I1227 08:05:04.603539 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c689587cc-6glpp" event={"ID":"72923f32-62d5-49f6-81dc-9741f36139e8","Type":"ContainerStarted","Data":"e7d6eff6886c52d929f057167f88c173d8c7bb889f8b427ba330a1d84183216a"} Dec 27 08:05:04 crc kubenswrapper[4934]: I1227 08:05:04.603810 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:05:04 crc kubenswrapper[4934]: I1227 08:05:04.605713 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-76g9s" event={"ID":"ea0027c6-81c4-4d65-96a6-43eb035d2043","Type":"ContainerStarted","Data":"416ae62c615270f259b28149ef9568a4f45c93acc55599fbdfd7470a080632e2"} Dec 27 08:05:04 crc kubenswrapper[4934]: I1227 08:05:04.606025 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:05:04 crc kubenswrapper[4934]: I1227 08:05:04.652972 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-76886c86d8-c6mdr" podStartSLOduration=9.652936364 podStartE2EDuration="9.652936364s" podCreationTimestamp="2025-12-27 08:04:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:05:04.641694187 +0000 UTC m=+1365.462134781" watchObservedRunningTime="2025-12-27 08:05:04.652936364 +0000 UTC m=+1365.473376998" Dec 27 08:05:04 crc kubenswrapper[4934]: I1227 08:05:04.670133 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-c689587cc-6glpp" podStartSLOduration=7.670070386 podStartE2EDuration="7.670070386s" podCreationTimestamp="2025-12-27 08:04:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:05:04.664270083 +0000 UTC m=+1365.484710677" watchObservedRunningTime="2025-12-27 08:05:04.670070386 +0000 UTC m=+1365.490511010" Dec 27 08:05:04 crc kubenswrapper[4934]: I1227 08:05:04.699912 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-76g9s" podStartSLOduration=9.699892851 podStartE2EDuration="9.699892851s" podCreationTimestamp="2025-12-27 08:04:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:05:04.682866052 +0000 UTC m=+1365.503306646" watchObservedRunningTime="2025-12-27 08:05:04.699892851 +0000 UTC m=+1365.520333445" Dec 27 08:05:05 crc kubenswrapper[4934]: I1227 08:05:05.621127 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-f745s" event={"ID":"5ef9a635-4d25-46d5-aa28-6c75f3cf8af4","Type":"ContainerStarted","Data":"f3f6afb85dd38d0d774bd8590ac099a7416e9ef26fa17b44420a18193ffcdcae"} Dec 27 08:05:05 crc kubenswrapper[4934]: I1227 08:05:05.659697 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-f745s" podStartSLOduration=2.5748726619999998 podStartE2EDuration="45.659665332s" podCreationTimestamp="2025-12-27 08:04:20 +0000 UTC" firstStartedPulling="2025-12-27 08:04:21.786895517 +0000 UTC m=+1322.607336121" lastFinishedPulling="2025-12-27 08:05:04.871688187 +0000 UTC m=+1365.692128791" observedRunningTime="2025-12-27 08:05:05.647654546 +0000 UTC m=+1366.468095210" watchObservedRunningTime="2025-12-27 08:05:05.659665332 +0000 UTC m=+1366.480105966" Dec 27 08:05:07 crc kubenswrapper[4934]: I1227 08:05:07.651105 4934 generic.go:334] "Generic (PLEG): container finished" podID="badf697c-e657-4b80-9483-63c94c87afaf" containerID="63ebcede11cc284f2621953206a69b0ab47a59fa73cafc0cd32b945c5b87bc0f" exitCode=0 Dec 27 08:05:07 crc kubenswrapper[4934]: I1227 08:05:07.651165 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"badf697c-e657-4b80-9483-63c94c87afaf","Type":"ContainerDied","Data":"63ebcede11cc284f2621953206a69b0ab47a59fa73cafc0cd32b945c5b87bc0f"} Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.034555 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.146453 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-config-data\") pod \"badf697c-e657-4b80-9483-63c94c87afaf\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.146572 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-sg-core-conf-yaml\") pod \"badf697c-e657-4b80-9483-63c94c87afaf\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.146681 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/badf697c-e657-4b80-9483-63c94c87afaf-log-httpd\") pod \"badf697c-e657-4b80-9483-63c94c87afaf\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.146764 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vksgf\" (UniqueName: \"kubernetes.io/projected/badf697c-e657-4b80-9483-63c94c87afaf-kube-api-access-vksgf\") pod \"badf697c-e657-4b80-9483-63c94c87afaf\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.146852 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/badf697c-e657-4b80-9483-63c94c87afaf-run-httpd\") pod \"badf697c-e657-4b80-9483-63c94c87afaf\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.147020 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-combined-ca-bundle\") pod \"badf697c-e657-4b80-9483-63c94c87afaf\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.147073 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-scripts\") pod \"badf697c-e657-4b80-9483-63c94c87afaf\" (UID: \"badf697c-e657-4b80-9483-63c94c87afaf\") " Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.148598 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/badf697c-e657-4b80-9483-63c94c87afaf-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "badf697c-e657-4b80-9483-63c94c87afaf" (UID: "badf697c-e657-4b80-9483-63c94c87afaf"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.148615 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/badf697c-e657-4b80-9483-63c94c87afaf-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "badf697c-e657-4b80-9483-63c94c87afaf" (UID: "badf697c-e657-4b80-9483-63c94c87afaf"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.155512 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/badf697c-e657-4b80-9483-63c94c87afaf-kube-api-access-vksgf" (OuterVolumeSpecName: "kube-api-access-vksgf") pod "badf697c-e657-4b80-9483-63c94c87afaf" (UID: "badf697c-e657-4b80-9483-63c94c87afaf"). InnerVolumeSpecName "kube-api-access-vksgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.155574 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-scripts" (OuterVolumeSpecName: "scripts") pod "badf697c-e657-4b80-9483-63c94c87afaf" (UID: "badf697c-e657-4b80-9483-63c94c87afaf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.160136 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "badf697c-e657-4b80-9483-63c94c87afaf" (UID: "badf697c-e657-4b80-9483-63c94c87afaf"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.250135 4934 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/badf697c-e657-4b80-9483-63c94c87afaf-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.250176 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vksgf\" (UniqueName: \"kubernetes.io/projected/badf697c-e657-4b80-9483-63c94c87afaf-kube-api-access-vksgf\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.250193 4934 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/badf697c-e657-4b80-9483-63c94c87afaf-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.250204 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.250219 4934 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.276552 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "badf697c-e657-4b80-9483-63c94c87afaf" (UID: "badf697c-e657-4b80-9483-63c94c87afaf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.287786 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-config-data" (OuterVolumeSpecName: "config-data") pod "badf697c-e657-4b80-9483-63c94c87afaf" (UID: "badf697c-e657-4b80-9483-63c94c87afaf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.352187 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.352216 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/badf697c-e657-4b80-9483-63c94c87afaf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.406958 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7fd8c76fd8-kbl58" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.687285 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"badf697c-e657-4b80-9483-63c94c87afaf","Type":"ContainerDied","Data":"27f762146bcb13869b79a8f4b3f2d96e8251bf8b9104e6f529417e628cdd07c6"} Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.687328 4934 scope.go:117] "RemoveContainer" containerID="533777f7714ac651df3ce7fadf06b65fc0d7096fe9f0cb914d969c67314f0142" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.687400 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.764888 4934 scope.go:117] "RemoveContainer" containerID="63ebcede11cc284f2621953206a69b0ab47a59fa73cafc0cd32b945c5b87bc0f" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.798630 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.810428 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.819157 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:05:09 crc kubenswrapper[4934]: E1227 08:05:09.822300 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="badf697c-e657-4b80-9483-63c94c87afaf" containerName="ceilometer-central-agent" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.822444 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="badf697c-e657-4b80-9483-63c94c87afaf" containerName="ceilometer-central-agent" Dec 27 08:05:09 crc kubenswrapper[4934]: E1227 08:05:09.822606 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="badf697c-e657-4b80-9483-63c94c87afaf" containerName="proxy-httpd" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.822624 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="badf697c-e657-4b80-9483-63c94c87afaf" containerName="proxy-httpd" Dec 27 08:05:09 crc kubenswrapper[4934]: E1227 08:05:09.822645 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="badf697c-e657-4b80-9483-63c94c87afaf" containerName="ceilometer-notification-agent" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.822652 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="badf697c-e657-4b80-9483-63c94c87afaf" containerName="ceilometer-notification-agent" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.823371 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="badf697c-e657-4b80-9483-63c94c87afaf" containerName="proxy-httpd" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.823848 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="badf697c-e657-4b80-9483-63c94c87afaf" containerName="ceilometer-notification-agent" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.823860 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="badf697c-e657-4b80-9483-63c94c87afaf" containerName="ceilometer-central-agent" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.829398 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.829499 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.831630 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.832217 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.926591 4934 scope.go:117] "RemoveContainer" containerID="45a559573157263c959004566fc36f27391dce93ff6361826ad2d9438ebd9a66" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.967922 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " pod="openstack/ceilometer-0" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.968308 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cef13c6c-b21d-45f8-83be-625000146643-run-httpd\") pod \"ceilometer-0\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " pod="openstack/ceilometer-0" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.968340 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cef13c6c-b21d-45f8-83be-625000146643-log-httpd\") pod \"ceilometer-0\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " pod="openstack/ceilometer-0" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.968365 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85ssg\" (UniqueName: \"kubernetes.io/projected/cef13c6c-b21d-45f8-83be-625000146643-kube-api-access-85ssg\") pod \"ceilometer-0\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " pod="openstack/ceilometer-0" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.968553 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-config-data\") pod \"ceilometer-0\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " pod="openstack/ceilometer-0" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.968653 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-scripts\") pod \"ceilometer-0\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " pod="openstack/ceilometer-0" Dec 27 08:05:09 crc kubenswrapper[4934]: I1227 08:05:09.968688 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " pod="openstack/ceilometer-0" Dec 27 08:05:10 crc kubenswrapper[4934]: I1227 08:05:10.070464 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cef13c6c-b21d-45f8-83be-625000146643-run-httpd\") pod \"ceilometer-0\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " pod="openstack/ceilometer-0" Dec 27 08:05:10 crc kubenswrapper[4934]: I1227 08:05:10.070532 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cef13c6c-b21d-45f8-83be-625000146643-log-httpd\") pod \"ceilometer-0\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " pod="openstack/ceilometer-0" Dec 27 08:05:10 crc kubenswrapper[4934]: I1227 08:05:10.070557 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85ssg\" (UniqueName: \"kubernetes.io/projected/cef13c6c-b21d-45f8-83be-625000146643-kube-api-access-85ssg\") pod \"ceilometer-0\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " pod="openstack/ceilometer-0" Dec 27 08:05:10 crc kubenswrapper[4934]: I1227 08:05:10.070640 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-config-data\") pod \"ceilometer-0\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " pod="openstack/ceilometer-0" Dec 27 08:05:10 crc kubenswrapper[4934]: I1227 08:05:10.070729 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-scripts\") pod \"ceilometer-0\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " pod="openstack/ceilometer-0" Dec 27 08:05:10 crc kubenswrapper[4934]: I1227 08:05:10.070773 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " pod="openstack/ceilometer-0" Dec 27 08:05:10 crc kubenswrapper[4934]: I1227 08:05:10.070814 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " pod="openstack/ceilometer-0" Dec 27 08:05:10 crc kubenswrapper[4934]: I1227 08:05:10.071879 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cef13c6c-b21d-45f8-83be-625000146643-run-httpd\") pod \"ceilometer-0\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " pod="openstack/ceilometer-0" Dec 27 08:05:10 crc kubenswrapper[4934]: I1227 08:05:10.075243 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cef13c6c-b21d-45f8-83be-625000146643-log-httpd\") pod \"ceilometer-0\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " pod="openstack/ceilometer-0" Dec 27 08:05:10 crc kubenswrapper[4934]: I1227 08:05:10.076691 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " pod="openstack/ceilometer-0" Dec 27 08:05:10 crc kubenswrapper[4934]: I1227 08:05:10.078579 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " pod="openstack/ceilometer-0" Dec 27 08:05:10 crc kubenswrapper[4934]: I1227 08:05:10.082992 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-config-data\") pod \"ceilometer-0\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " pod="openstack/ceilometer-0" Dec 27 08:05:10 crc kubenswrapper[4934]: I1227 08:05:10.083385 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-scripts\") pod \"ceilometer-0\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " pod="openstack/ceilometer-0" Dec 27 08:05:10 crc kubenswrapper[4934]: I1227 08:05:10.091002 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85ssg\" (UniqueName: \"kubernetes.io/projected/cef13c6c-b21d-45f8-83be-625000146643-kube-api-access-85ssg\") pod \"ceilometer-0\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " pod="openstack/ceilometer-0" Dec 27 08:05:10 crc kubenswrapper[4934]: I1227 08:05:10.160641 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:05:10 crc kubenswrapper[4934]: I1227 08:05:10.731903 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:05:10 crc kubenswrapper[4934]: W1227 08:05:10.739064 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcef13c6c_b21d_45f8_83be_625000146643.slice/crio-3ed37090c304cdcaebe47b77760b5dcdd36d526383ff2393e1885a17c60944b6 WatchSource:0}: Error finding container 3ed37090c304cdcaebe47b77760b5dcdd36d526383ff2393e1885a17c60944b6: Status 404 returned error can't find the container with id 3ed37090c304cdcaebe47b77760b5dcdd36d526383ff2393e1885a17c60944b6 Dec 27 08:05:11 crc kubenswrapper[4934]: I1227 08:05:11.170290 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:05:11 crc kubenswrapper[4934]: I1227 08:05:11.229862 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-dw8tc"] Dec 27 08:05:11 crc kubenswrapper[4934]: I1227 08:05:11.230102 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" podUID="1df64eee-76a8-4be2-8b36-665bc0f1564f" containerName="dnsmasq-dns" containerID="cri-o://d3369af297081db2f9235d0094a43a80f5646ff1b6ceca3e41271bb4dc40ab4f" gracePeriod=10 Dec 27 08:05:11 crc kubenswrapper[4934]: I1227 08:05:11.480554 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="badf697c-e657-4b80-9483-63c94c87afaf" path="/var/lib/kubelet/pods/badf697c-e657-4b80-9483-63c94c87afaf/volumes" Dec 27 08:05:11 crc kubenswrapper[4934]: I1227 08:05:11.715196 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rrj89" event={"ID":"72d0ed22-31be-4048-b031-139856c584c5","Type":"ContainerStarted","Data":"dd3aa6e013a5c434c3c2c70132d259a4f0966aee7c14cdfa3dc77dd5572fc961"} Dec 27 08:05:11 crc kubenswrapper[4934]: I1227 08:05:11.717797 4934 generic.go:334] "Generic (PLEG): container finished" podID="1df64eee-76a8-4be2-8b36-665bc0f1564f" containerID="d3369af297081db2f9235d0094a43a80f5646ff1b6ceca3e41271bb4dc40ab4f" exitCode=0 Dec 27 08:05:11 crc kubenswrapper[4934]: I1227 08:05:11.717839 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" event={"ID":"1df64eee-76a8-4be2-8b36-665bc0f1564f","Type":"ContainerDied","Data":"d3369af297081db2f9235d0094a43a80f5646ff1b6ceca3e41271bb4dc40ab4f"} Dec 27 08:05:11 crc kubenswrapper[4934]: I1227 08:05:11.719450 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cef13c6c-b21d-45f8-83be-625000146643","Type":"ContainerStarted","Data":"3ed37090c304cdcaebe47b77760b5dcdd36d526383ff2393e1885a17c60944b6"} Dec 27 08:05:11 crc kubenswrapper[4934]: I1227 08:05:11.742770 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-rrj89" podStartSLOduration=4.27289477 podStartE2EDuration="52.742750292s" podCreationTimestamp="2025-12-27 08:04:19 +0000 UTC" firstStartedPulling="2025-12-27 08:04:21.787272587 +0000 UTC m=+1322.607713181" lastFinishedPulling="2025-12-27 08:05:10.257128109 +0000 UTC m=+1371.077568703" observedRunningTime="2025-12-27 08:05:11.739998004 +0000 UTC m=+1372.560438598" watchObservedRunningTime="2025-12-27 08:05:11.742750292 +0000 UTC m=+1372.563190886" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.307896 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.312879 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.315438 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.318203 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-tm24t" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.318466 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.360888 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.371707 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.455452 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-657bv\" (UniqueName: \"kubernetes.io/projected/1df64eee-76a8-4be2-8b36-665bc0f1564f-kube-api-access-657bv\") pod \"1df64eee-76a8-4be2-8b36-665bc0f1564f\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.455514 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-dns-svc\") pod \"1df64eee-76a8-4be2-8b36-665bc0f1564f\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.455605 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-dns-swift-storage-0\") pod \"1df64eee-76a8-4be2-8b36-665bc0f1564f\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.455628 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-ovsdbserver-sb\") pod \"1df64eee-76a8-4be2-8b36-665bc0f1564f\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.455723 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-ovsdbserver-nb\") pod \"1df64eee-76a8-4be2-8b36-665bc0f1564f\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.455779 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-config\") pod \"1df64eee-76a8-4be2-8b36-665bc0f1564f\" (UID: \"1df64eee-76a8-4be2-8b36-665bc0f1564f\") " Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.456097 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f5eabd4f-a103-4ec0-9c14-7deb0e36fe61-openstack-config\") pod \"openstackclient\" (UID: \"f5eabd4f-a103-4ec0-9c14-7deb0e36fe61\") " pod="openstack/openstackclient" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.456133 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7qpf\" (UniqueName: \"kubernetes.io/projected/f5eabd4f-a103-4ec0-9c14-7deb0e36fe61-kube-api-access-d7qpf\") pod \"openstackclient\" (UID: \"f5eabd4f-a103-4ec0-9c14-7deb0e36fe61\") " pod="openstack/openstackclient" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.456263 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f5eabd4f-a103-4ec0-9c14-7deb0e36fe61-openstack-config-secret\") pod \"openstackclient\" (UID: \"f5eabd4f-a103-4ec0-9c14-7deb0e36fe61\") " pod="openstack/openstackclient" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.456307 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5eabd4f-a103-4ec0-9c14-7deb0e36fe61-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f5eabd4f-a103-4ec0-9c14-7deb0e36fe61\") " pod="openstack/openstackclient" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.477950 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1df64eee-76a8-4be2-8b36-665bc0f1564f-kube-api-access-657bv" (OuterVolumeSpecName: "kube-api-access-657bv") pod "1df64eee-76a8-4be2-8b36-665bc0f1564f" (UID: "1df64eee-76a8-4be2-8b36-665bc0f1564f"). InnerVolumeSpecName "kube-api-access-657bv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.525942 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1df64eee-76a8-4be2-8b36-665bc0f1564f" (UID: "1df64eee-76a8-4be2-8b36-665bc0f1564f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.538263 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1df64eee-76a8-4be2-8b36-665bc0f1564f" (UID: "1df64eee-76a8-4be2-8b36-665bc0f1564f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.540733 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1df64eee-76a8-4be2-8b36-665bc0f1564f" (UID: "1df64eee-76a8-4be2-8b36-665bc0f1564f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.547260 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-config" (OuterVolumeSpecName: "config") pod "1df64eee-76a8-4be2-8b36-665bc0f1564f" (UID: "1df64eee-76a8-4be2-8b36-665bc0f1564f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.556334 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1df64eee-76a8-4be2-8b36-665bc0f1564f" (UID: "1df64eee-76a8-4be2-8b36-665bc0f1564f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.557688 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f5eabd4f-a103-4ec0-9c14-7deb0e36fe61-openstack-config-secret\") pod \"openstackclient\" (UID: \"f5eabd4f-a103-4ec0-9c14-7deb0e36fe61\") " pod="openstack/openstackclient" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.557739 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5eabd4f-a103-4ec0-9c14-7deb0e36fe61-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f5eabd4f-a103-4ec0-9c14-7deb0e36fe61\") " pod="openstack/openstackclient" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.557817 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f5eabd4f-a103-4ec0-9c14-7deb0e36fe61-openstack-config\") pod \"openstackclient\" (UID: \"f5eabd4f-a103-4ec0-9c14-7deb0e36fe61\") " pod="openstack/openstackclient" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.557847 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7qpf\" (UniqueName: \"kubernetes.io/projected/f5eabd4f-a103-4ec0-9c14-7deb0e36fe61-kube-api-access-d7qpf\") pod \"openstackclient\" (UID: \"f5eabd4f-a103-4ec0-9c14-7deb0e36fe61\") " pod="openstack/openstackclient" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.557924 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-657bv\" (UniqueName: \"kubernetes.io/projected/1df64eee-76a8-4be2-8b36-665bc0f1564f-kube-api-access-657bv\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.557939 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.557949 4934 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.557957 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.557967 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.557978 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1df64eee-76a8-4be2-8b36-665bc0f1564f-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.559010 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/f5eabd4f-a103-4ec0-9c14-7deb0e36fe61-openstack-config\") pod \"openstackclient\" (UID: \"f5eabd4f-a103-4ec0-9c14-7deb0e36fe61\") " pod="openstack/openstackclient" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.560830 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f5eabd4f-a103-4ec0-9c14-7deb0e36fe61-openstack-config-secret\") pod \"openstackclient\" (UID: \"f5eabd4f-a103-4ec0-9c14-7deb0e36fe61\") " pod="openstack/openstackclient" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.562044 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5eabd4f-a103-4ec0-9c14-7deb0e36fe61-combined-ca-bundle\") pod \"openstackclient\" (UID: \"f5eabd4f-a103-4ec0-9c14-7deb0e36fe61\") " pod="openstack/openstackclient" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.575831 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7qpf\" (UniqueName: \"kubernetes.io/projected/f5eabd4f-a103-4ec0-9c14-7deb0e36fe61-kube-api-access-d7qpf\") pod \"openstackclient\" (UID: \"f5eabd4f-a103-4ec0-9c14-7deb0e36fe61\") " pod="openstack/openstackclient" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.686496 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.747487 4934 generic.go:334] "Generic (PLEG): container finished" podID="61f5132a-303d-4f76-ae58-c57d82a59fd4" containerID="db4ed2f28d2e711d823b1ea76d20bc93ee4ecdf738544c01eaea7df7995d7b4a" exitCode=0 Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.747552 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-5qllw" event={"ID":"61f5132a-303d-4f76-ae58-c57d82a59fd4","Type":"ContainerDied","Data":"db4ed2f28d2e711d823b1ea76d20bc93ee4ecdf738544c01eaea7df7995d7b4a"} Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.749427 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" event={"ID":"1df64eee-76a8-4be2-8b36-665bc0f1564f","Type":"ContainerDied","Data":"895bd359c1243b6a62cfe2d688eb1f44e222bf71ad62c64a706630b35365899c"} Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.749433 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-dw8tc" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.749460 4934 scope.go:117] "RemoveContainer" containerID="d3369af297081db2f9235d0094a43a80f5646ff1b6ceca3e41271bb4dc40ab4f" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.753708 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cef13c6c-b21d-45f8-83be-625000146643","Type":"ContainerStarted","Data":"ea63ad9ef292b297cb30f1be32c7e34a88199aa88c4d09e99e027b160549fcb0"} Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.799536 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-dw8tc"] Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.800101 4934 scope.go:117] "RemoveContainer" containerID="0a31e676527400f5082b959a90c929559a51b09287f384910eef01bfcc64300d" Dec 27 08:05:13 crc kubenswrapper[4934]: I1227 08:05:13.812430 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-dw8tc"] Dec 27 08:05:14 crc kubenswrapper[4934]: I1227 08:05:14.220639 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 27 08:05:14 crc kubenswrapper[4934]: I1227 08:05:14.766989 4934 generic.go:334] "Generic (PLEG): container finished" podID="5ef9a635-4d25-46d5-aa28-6c75f3cf8af4" containerID="f3f6afb85dd38d0d774bd8590ac099a7416e9ef26fa17b44420a18193ffcdcae" exitCode=0 Dec 27 08:05:14 crc kubenswrapper[4934]: I1227 08:05:14.767117 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-f745s" event={"ID":"5ef9a635-4d25-46d5-aa28-6c75f3cf8af4","Type":"ContainerDied","Data":"f3f6afb85dd38d0d774bd8590ac099a7416e9ef26fa17b44420a18193ffcdcae"} Dec 27 08:05:14 crc kubenswrapper[4934]: I1227 08:05:14.769255 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f5eabd4f-a103-4ec0-9c14-7deb0e36fe61","Type":"ContainerStarted","Data":"b9cca65d6ae68bd740c2d18fe7356a97d5c25c23543f18ae35f0d1718b2d0c11"} Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.256415 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-5qllw" Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.330070 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.330149 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.330245 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.331602 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b5285e4beb4b025eb5cca462bbc9a5da90660cedc2a9f99c3f12f8c689605503"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.331671 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://b5285e4beb4b025eb5cca462bbc9a5da90660cedc2a9f99c3f12f8c689605503" gracePeriod=600 Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.401669 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61f5132a-303d-4f76-ae58-c57d82a59fd4-config-data\") pod \"61f5132a-303d-4f76-ae58-c57d82a59fd4\" (UID: \"61f5132a-303d-4f76-ae58-c57d82a59fd4\") " Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.401759 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9s9pr\" (UniqueName: \"kubernetes.io/projected/61f5132a-303d-4f76-ae58-c57d82a59fd4-kube-api-access-9s9pr\") pod \"61f5132a-303d-4f76-ae58-c57d82a59fd4\" (UID: \"61f5132a-303d-4f76-ae58-c57d82a59fd4\") " Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.402066 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61f5132a-303d-4f76-ae58-c57d82a59fd4-combined-ca-bundle\") pod \"61f5132a-303d-4f76-ae58-c57d82a59fd4\" (UID: \"61f5132a-303d-4f76-ae58-c57d82a59fd4\") " Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.432322 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61f5132a-303d-4f76-ae58-c57d82a59fd4-kube-api-access-9s9pr" (OuterVolumeSpecName: "kube-api-access-9s9pr") pod "61f5132a-303d-4f76-ae58-c57d82a59fd4" (UID: "61f5132a-303d-4f76-ae58-c57d82a59fd4"). InnerVolumeSpecName "kube-api-access-9s9pr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.435191 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61f5132a-303d-4f76-ae58-c57d82a59fd4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "61f5132a-303d-4f76-ae58-c57d82a59fd4" (UID: "61f5132a-303d-4f76-ae58-c57d82a59fd4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.484037 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1df64eee-76a8-4be2-8b36-665bc0f1564f" path="/var/lib/kubelet/pods/1df64eee-76a8-4be2-8b36-665bc0f1564f/volumes" Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.495728 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61f5132a-303d-4f76-ae58-c57d82a59fd4-config-data" (OuterVolumeSpecName: "config-data") pod "61f5132a-303d-4f76-ae58-c57d82a59fd4" (UID: "61f5132a-303d-4f76-ae58-c57d82a59fd4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.508193 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/61f5132a-303d-4f76-ae58-c57d82a59fd4-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.508226 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9s9pr\" (UniqueName: \"kubernetes.io/projected/61f5132a-303d-4f76-ae58-c57d82a59fd4-kube-api-access-9s9pr\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.508237 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61f5132a-303d-4f76-ae58-c57d82a59fd4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.784974 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-5qllw" event={"ID":"61f5132a-303d-4f76-ae58-c57d82a59fd4","Type":"ContainerDied","Data":"75b225b078c8d1ec9d82022861232f98fd169bb3b1758cfaac38f3787e83f71a"} Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.784998 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-5qllw" Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.785022 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75b225b078c8d1ec9d82022861232f98fd169bb3b1758cfaac38f3787e83f71a" Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.787543 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cef13c6c-b21d-45f8-83be-625000146643","Type":"ContainerStarted","Data":"157284b0f56ac0f09ace1cb65b4cc1ae55750e564536da9dd030cd2fed0a61da"} Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.806328 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="b5285e4beb4b025eb5cca462bbc9a5da90660cedc2a9f99c3f12f8c689605503" exitCode=0 Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.806561 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"b5285e4beb4b025eb5cca462bbc9a5da90660cedc2a9f99c3f12f8c689605503"} Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.806588 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab"} Dec 27 08:05:15 crc kubenswrapper[4934]: I1227 08:05:15.806602 4934 scope.go:117] "RemoveContainer" containerID="b960fbd3418f51dd41277ee06c40ec3eb700b9243e777f61c98b4b5b56bf0c98" Dec 27 08:05:16 crc kubenswrapper[4934]: I1227 08:05:16.296924 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-f745s" Dec 27 08:05:16 crc kubenswrapper[4934]: I1227 08:05:16.430912 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5ef9a635-4d25-46d5-aa28-6c75f3cf8af4-db-sync-config-data\") pod \"5ef9a635-4d25-46d5-aa28-6c75f3cf8af4\" (UID: \"5ef9a635-4d25-46d5-aa28-6c75f3cf8af4\") " Dec 27 08:05:16 crc kubenswrapper[4934]: I1227 08:05:16.431002 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfm8h\" (UniqueName: \"kubernetes.io/projected/5ef9a635-4d25-46d5-aa28-6c75f3cf8af4-kube-api-access-wfm8h\") pod \"5ef9a635-4d25-46d5-aa28-6c75f3cf8af4\" (UID: \"5ef9a635-4d25-46d5-aa28-6c75f3cf8af4\") " Dec 27 08:05:16 crc kubenswrapper[4934]: I1227 08:05:16.431132 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ef9a635-4d25-46d5-aa28-6c75f3cf8af4-combined-ca-bundle\") pod \"5ef9a635-4d25-46d5-aa28-6c75f3cf8af4\" (UID: \"5ef9a635-4d25-46d5-aa28-6c75f3cf8af4\") " Dec 27 08:05:16 crc kubenswrapper[4934]: I1227 08:05:16.461011 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ef9a635-4d25-46d5-aa28-6c75f3cf8af4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "5ef9a635-4d25-46d5-aa28-6c75f3cf8af4" (UID: "5ef9a635-4d25-46d5-aa28-6c75f3cf8af4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:16 crc kubenswrapper[4934]: I1227 08:05:16.468455 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ef9a635-4d25-46d5-aa28-6c75f3cf8af4-kube-api-access-wfm8h" (OuterVolumeSpecName: "kube-api-access-wfm8h") pod "5ef9a635-4d25-46d5-aa28-6c75f3cf8af4" (UID: "5ef9a635-4d25-46d5-aa28-6c75f3cf8af4"). InnerVolumeSpecName "kube-api-access-wfm8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:16 crc kubenswrapper[4934]: I1227 08:05:16.531253 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ef9a635-4d25-46d5-aa28-6c75f3cf8af4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ef9a635-4d25-46d5-aa28-6c75f3cf8af4" (UID: "5ef9a635-4d25-46d5-aa28-6c75f3cf8af4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:16 crc kubenswrapper[4934]: I1227 08:05:16.533810 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ef9a635-4d25-46d5-aa28-6c75f3cf8af4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:16 crc kubenswrapper[4934]: I1227 08:05:16.533846 4934 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5ef9a635-4d25-46d5-aa28-6c75f3cf8af4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:16 crc kubenswrapper[4934]: I1227 08:05:16.533857 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfm8h\" (UniqueName: \"kubernetes.io/projected/5ef9a635-4d25-46d5-aa28-6c75f3cf8af4-kube-api-access-wfm8h\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:16 crc kubenswrapper[4934]: I1227 08:05:16.822156 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-f745s" event={"ID":"5ef9a635-4d25-46d5-aa28-6c75f3cf8af4","Type":"ContainerDied","Data":"b65dc62674e3ade81c1c692855a0e1e48d48174edf673179ad47a36f0dba0e2e"} Dec 27 08:05:16 crc kubenswrapper[4934]: I1227 08:05:16.822189 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b65dc62674e3ade81c1c692855a0e1e48d48174edf673179ad47a36f0dba0e2e" Dec 27 08:05:16 crc kubenswrapper[4934]: I1227 08:05:16.822248 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-f745s" Dec 27 08:05:16 crc kubenswrapper[4934]: I1227 08:05:16.828395 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cef13c6c-b21d-45f8-83be-625000146643","Type":"ContainerStarted","Data":"0e4457b0013fbecad3139a089d129032412a51cceb7d10b391d48c65f4f84f4f"} Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.039691 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-859445d775-g44pr"] Dec 27 08:05:17 crc kubenswrapper[4934]: E1227 08:05:17.040878 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ef9a635-4d25-46d5-aa28-6c75f3cf8af4" containerName="barbican-db-sync" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.040895 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ef9a635-4d25-46d5-aa28-6c75f3cf8af4" containerName="barbican-db-sync" Dec 27 08:05:17 crc kubenswrapper[4934]: E1227 08:05:17.040922 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61f5132a-303d-4f76-ae58-c57d82a59fd4" containerName="heat-db-sync" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.040928 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="61f5132a-303d-4f76-ae58-c57d82a59fd4" containerName="heat-db-sync" Dec 27 08:05:17 crc kubenswrapper[4934]: E1227 08:05:17.040943 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df64eee-76a8-4be2-8b36-665bc0f1564f" containerName="dnsmasq-dns" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.040949 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df64eee-76a8-4be2-8b36-665bc0f1564f" containerName="dnsmasq-dns" Dec 27 08:05:17 crc kubenswrapper[4934]: E1227 08:05:17.040962 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1df64eee-76a8-4be2-8b36-665bc0f1564f" containerName="init" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.040967 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1df64eee-76a8-4be2-8b36-665bc0f1564f" containerName="init" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.041160 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ef9a635-4d25-46d5-aa28-6c75f3cf8af4" containerName="barbican-db-sync" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.041174 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1df64eee-76a8-4be2-8b36-665bc0f1564f" containerName="dnsmasq-dns" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.041194 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="61f5132a-303d-4f76-ae58-c57d82a59fd4" containerName="heat-db-sync" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.042247 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-859445d775-g44pr" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.046689 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-bx627" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.047050 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.047247 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.083767 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-859445d775-g44pr"] Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.117125 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-654859b66d-km642"] Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.133566 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-654859b66d-km642" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.139947 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-654859b66d-km642"] Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.151293 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.155185 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f083ef4-9928-400b-969e-1df70b3e9475-logs\") pod \"barbican-worker-859445d775-g44pr\" (UID: \"3f083ef4-9928-400b-969e-1df70b3e9475\") " pod="openstack/barbican-worker-859445d775-g44pr" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.155306 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f083ef4-9928-400b-969e-1df70b3e9475-config-data-custom\") pod \"barbican-worker-859445d775-g44pr\" (UID: \"3f083ef4-9928-400b-969e-1df70b3e9475\") " pod="openstack/barbican-worker-859445d775-g44pr" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.155368 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f083ef4-9928-400b-969e-1df70b3e9475-config-data\") pod \"barbican-worker-859445d775-g44pr\" (UID: \"3f083ef4-9928-400b-969e-1df70b3e9475\") " pod="openstack/barbican-worker-859445d775-g44pr" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.155445 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f083ef4-9928-400b-969e-1df70b3e9475-combined-ca-bundle\") pod \"barbican-worker-859445d775-g44pr\" (UID: \"3f083ef4-9928-400b-969e-1df70b3e9475\") " pod="openstack/barbican-worker-859445d775-g44pr" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.155573 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m9qf\" (UniqueName: \"kubernetes.io/projected/3f083ef4-9928-400b-969e-1df70b3e9475-kube-api-access-4m9qf\") pod \"barbican-worker-859445d775-g44pr\" (UID: \"3f083ef4-9928-400b-969e-1df70b3e9475\") " pod="openstack/barbican-worker-859445d775-g44pr" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.170386 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-xlm2k"] Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.172722 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.224075 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-xlm2k"] Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.265691 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m9qf\" (UniqueName: \"kubernetes.io/projected/3f083ef4-9928-400b-969e-1df70b3e9475-kube-api-access-4m9qf\") pod \"barbican-worker-859445d775-g44pr\" (UID: \"3f083ef4-9928-400b-969e-1df70b3e9475\") " pod="openstack/barbican-worker-859445d775-g44pr" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.265739 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700034e0-f221-4bca-a973-a6d49c7dc406-combined-ca-bundle\") pod \"barbican-keystone-listener-654859b66d-km642\" (UID: \"700034e0-f221-4bca-a973-a6d49c7dc406\") " pod="openstack/barbican-keystone-listener-654859b66d-km642" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.265774 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-config\") pod \"dnsmasq-dns-848cf88cfc-xlm2k\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.265797 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700034e0-f221-4bca-a973-a6d49c7dc406-logs\") pod \"barbican-keystone-listener-654859b66d-km642\" (UID: \"700034e0-f221-4bca-a973-a6d49c7dc406\") " pod="openstack/barbican-keystone-listener-654859b66d-km642" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.265845 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f083ef4-9928-400b-969e-1df70b3e9475-logs\") pod \"barbican-worker-859445d775-g44pr\" (UID: \"3f083ef4-9928-400b-969e-1df70b3e9475\") " pod="openstack/barbican-worker-859445d775-g44pr" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.265869 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-xlm2k\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.265897 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-xlm2k\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.265914 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdjlf\" (UniqueName: \"kubernetes.io/projected/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-kube-api-access-fdjlf\") pod \"dnsmasq-dns-848cf88cfc-xlm2k\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.265937 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-xlm2k\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.265956 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/700034e0-f221-4bca-a973-a6d49c7dc406-config-data-custom\") pod \"barbican-keystone-listener-654859b66d-km642\" (UID: \"700034e0-f221-4bca-a973-a6d49c7dc406\") " pod="openstack/barbican-keystone-listener-654859b66d-km642" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.265973 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqt8z\" (UniqueName: \"kubernetes.io/projected/700034e0-f221-4bca-a973-a6d49c7dc406-kube-api-access-rqt8z\") pod \"barbican-keystone-listener-654859b66d-km642\" (UID: \"700034e0-f221-4bca-a973-a6d49c7dc406\") " pod="openstack/barbican-keystone-listener-654859b66d-km642" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.265991 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f083ef4-9928-400b-969e-1df70b3e9475-config-data-custom\") pod \"barbican-worker-859445d775-g44pr\" (UID: \"3f083ef4-9928-400b-969e-1df70b3e9475\") " pod="openstack/barbican-worker-859445d775-g44pr" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.266025 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f083ef4-9928-400b-969e-1df70b3e9475-config-data\") pod \"barbican-worker-859445d775-g44pr\" (UID: \"3f083ef4-9928-400b-969e-1df70b3e9475\") " pod="openstack/barbican-worker-859445d775-g44pr" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.266050 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700034e0-f221-4bca-a973-a6d49c7dc406-config-data\") pod \"barbican-keystone-listener-654859b66d-km642\" (UID: \"700034e0-f221-4bca-a973-a6d49c7dc406\") " pod="openstack/barbican-keystone-listener-654859b66d-km642" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.266094 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f083ef4-9928-400b-969e-1df70b3e9475-combined-ca-bundle\") pod \"barbican-worker-859445d775-g44pr\" (UID: \"3f083ef4-9928-400b-969e-1df70b3e9475\") " pod="openstack/barbican-worker-859445d775-g44pr" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.266124 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-xlm2k\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.267999 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f083ef4-9928-400b-969e-1df70b3e9475-logs\") pod \"barbican-worker-859445d775-g44pr\" (UID: \"3f083ef4-9928-400b-969e-1df70b3e9475\") " pod="openstack/barbican-worker-859445d775-g44pr" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.292140 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f083ef4-9928-400b-969e-1df70b3e9475-config-data\") pod \"barbican-worker-859445d775-g44pr\" (UID: \"3f083ef4-9928-400b-969e-1df70b3e9475\") " pod="openstack/barbican-worker-859445d775-g44pr" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.292459 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-58d7546b8b-lqdtc"] Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.292886 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f083ef4-9928-400b-969e-1df70b3e9475-combined-ca-bundle\") pod \"barbican-worker-859445d775-g44pr\" (UID: \"3f083ef4-9928-400b-969e-1df70b3e9475\") " pod="openstack/barbican-worker-859445d775-g44pr" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.293842 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f083ef4-9928-400b-969e-1df70b3e9475-config-data-custom\") pod \"barbican-worker-859445d775-g44pr\" (UID: \"3f083ef4-9928-400b-969e-1df70b3e9475\") " pod="openstack/barbican-worker-859445d775-g44pr" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.294286 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.298236 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m9qf\" (UniqueName: \"kubernetes.io/projected/3f083ef4-9928-400b-969e-1df70b3e9475-kube-api-access-4m9qf\") pod \"barbican-worker-859445d775-g44pr\" (UID: \"3f083ef4-9928-400b-969e-1df70b3e9475\") " pod="openstack/barbican-worker-859445d775-g44pr" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.299274 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.358118 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-58d7546b8b-lqdtc"] Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.367542 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700034e0-f221-4bca-a973-a6d49c7dc406-combined-ca-bundle\") pod \"barbican-keystone-listener-654859b66d-km642\" (UID: \"700034e0-f221-4bca-a973-a6d49c7dc406\") " pod="openstack/barbican-keystone-listener-654859b66d-km642" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.367615 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-config\") pod \"dnsmasq-dns-848cf88cfc-xlm2k\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.367646 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700034e0-f221-4bca-a973-a6d49c7dc406-logs\") pod \"barbican-keystone-listener-654859b66d-km642\" (UID: \"700034e0-f221-4bca-a973-a6d49c7dc406\") " pod="openstack/barbican-keystone-listener-654859b66d-km642" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.367670 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/327eb20e-e905-48a6-99b8-9209813e3330-config-data-custom\") pod \"barbican-api-58d7546b8b-lqdtc\" (UID: \"327eb20e-e905-48a6-99b8-9209813e3330\") " pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.367700 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/327eb20e-e905-48a6-99b8-9209813e3330-config-data\") pod \"barbican-api-58d7546b8b-lqdtc\" (UID: \"327eb20e-e905-48a6-99b8-9209813e3330\") " pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.367752 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-xlm2k\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.367777 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/327eb20e-e905-48a6-99b8-9209813e3330-logs\") pod \"barbican-api-58d7546b8b-lqdtc\" (UID: \"327eb20e-e905-48a6-99b8-9209813e3330\") " pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.367801 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-xlm2k\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.367818 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdjlf\" (UniqueName: \"kubernetes.io/projected/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-kube-api-access-fdjlf\") pod \"dnsmasq-dns-848cf88cfc-xlm2k\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.367842 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-xlm2k\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.367862 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/700034e0-f221-4bca-a973-a6d49c7dc406-config-data-custom\") pod \"barbican-keystone-listener-654859b66d-km642\" (UID: \"700034e0-f221-4bca-a973-a6d49c7dc406\") " pod="openstack/barbican-keystone-listener-654859b66d-km642" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.367882 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqt8z\" (UniqueName: \"kubernetes.io/projected/700034e0-f221-4bca-a973-a6d49c7dc406-kube-api-access-rqt8z\") pod \"barbican-keystone-listener-654859b66d-km642\" (UID: \"700034e0-f221-4bca-a973-a6d49c7dc406\") " pod="openstack/barbican-keystone-listener-654859b66d-km642" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.368164 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700034e0-f221-4bca-a973-a6d49c7dc406-config-data\") pod \"barbican-keystone-listener-654859b66d-km642\" (UID: \"700034e0-f221-4bca-a973-a6d49c7dc406\") " pod="openstack/barbican-keystone-listener-654859b66d-km642" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.368320 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/327eb20e-e905-48a6-99b8-9209813e3330-combined-ca-bundle\") pod \"barbican-api-58d7546b8b-lqdtc\" (UID: \"327eb20e-e905-48a6-99b8-9209813e3330\") " pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.368340 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw5vp\" (UniqueName: \"kubernetes.io/projected/327eb20e-e905-48a6-99b8-9209813e3330-kube-api-access-fw5vp\") pod \"barbican-api-58d7546b8b-lqdtc\" (UID: \"327eb20e-e905-48a6-99b8-9209813e3330\") " pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.368362 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-xlm2k\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.368496 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/700034e0-f221-4bca-a973-a6d49c7dc406-logs\") pod \"barbican-keystone-listener-654859b66d-km642\" (UID: \"700034e0-f221-4bca-a973-a6d49c7dc406\") " pod="openstack/barbican-keystone-listener-654859b66d-km642" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.369676 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-config\") pod \"dnsmasq-dns-848cf88cfc-xlm2k\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.370186 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-xlm2k\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.370257 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-xlm2k\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.371204 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-xlm2k\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.371315 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-xlm2k\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.371952 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/700034e0-f221-4bca-a973-a6d49c7dc406-combined-ca-bundle\") pod \"barbican-keystone-listener-654859b66d-km642\" (UID: \"700034e0-f221-4bca-a973-a6d49c7dc406\") " pod="openstack/barbican-keystone-listener-654859b66d-km642" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.374983 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/700034e0-f221-4bca-a973-a6d49c7dc406-config-data-custom\") pod \"barbican-keystone-listener-654859b66d-km642\" (UID: \"700034e0-f221-4bca-a973-a6d49c7dc406\") " pod="openstack/barbican-keystone-listener-654859b66d-km642" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.375522 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-859445d775-g44pr" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.376890 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/700034e0-f221-4bca-a973-a6d49c7dc406-config-data\") pod \"barbican-keystone-listener-654859b66d-km642\" (UID: \"700034e0-f221-4bca-a973-a6d49c7dc406\") " pod="openstack/barbican-keystone-listener-654859b66d-km642" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.396387 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdjlf\" (UniqueName: \"kubernetes.io/projected/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-kube-api-access-fdjlf\") pod \"dnsmasq-dns-848cf88cfc-xlm2k\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.398110 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqt8z\" (UniqueName: \"kubernetes.io/projected/700034e0-f221-4bca-a973-a6d49c7dc406-kube-api-access-rqt8z\") pod \"barbican-keystone-listener-654859b66d-km642\" (UID: \"700034e0-f221-4bca-a973-a6d49c7dc406\") " pod="openstack/barbican-keystone-listener-654859b66d-km642" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.470802 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/327eb20e-e905-48a6-99b8-9209813e3330-combined-ca-bundle\") pod \"barbican-api-58d7546b8b-lqdtc\" (UID: \"327eb20e-e905-48a6-99b8-9209813e3330\") " pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.471054 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw5vp\" (UniqueName: \"kubernetes.io/projected/327eb20e-e905-48a6-99b8-9209813e3330-kube-api-access-fw5vp\") pod \"barbican-api-58d7546b8b-lqdtc\" (UID: \"327eb20e-e905-48a6-99b8-9209813e3330\") " pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.471151 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/327eb20e-e905-48a6-99b8-9209813e3330-config-data-custom\") pod \"barbican-api-58d7546b8b-lqdtc\" (UID: \"327eb20e-e905-48a6-99b8-9209813e3330\") " pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.471196 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/327eb20e-e905-48a6-99b8-9209813e3330-config-data\") pod \"barbican-api-58d7546b8b-lqdtc\" (UID: \"327eb20e-e905-48a6-99b8-9209813e3330\") " pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.471257 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/327eb20e-e905-48a6-99b8-9209813e3330-logs\") pod \"barbican-api-58d7546b8b-lqdtc\" (UID: \"327eb20e-e905-48a6-99b8-9209813e3330\") " pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.471921 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/327eb20e-e905-48a6-99b8-9209813e3330-logs\") pod \"barbican-api-58d7546b8b-lqdtc\" (UID: \"327eb20e-e905-48a6-99b8-9209813e3330\") " pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.476558 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/327eb20e-e905-48a6-99b8-9209813e3330-combined-ca-bundle\") pod \"barbican-api-58d7546b8b-lqdtc\" (UID: \"327eb20e-e905-48a6-99b8-9209813e3330\") " pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.483196 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/327eb20e-e905-48a6-99b8-9209813e3330-config-data-custom\") pod \"barbican-api-58d7546b8b-lqdtc\" (UID: \"327eb20e-e905-48a6-99b8-9209813e3330\") " pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.483936 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/327eb20e-e905-48a6-99b8-9209813e3330-config-data\") pod \"barbican-api-58d7546b8b-lqdtc\" (UID: \"327eb20e-e905-48a6-99b8-9209813e3330\") " pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.492481 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-654859b66d-km642" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.501756 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw5vp\" (UniqueName: \"kubernetes.io/projected/327eb20e-e905-48a6-99b8-9209813e3330-kube-api-access-fw5vp\") pod \"barbican-api-58d7546b8b-lqdtc\" (UID: \"327eb20e-e905-48a6-99b8-9209813e3330\") " pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.515692 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.527861 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.901346 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cef13c6c-b21d-45f8-83be-625000146643","Type":"ContainerStarted","Data":"75b87c79a50fe702df30afcea7ea1735e5bab70d914b25d800a4bf0503db1a01"} Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.902025 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 27 08:05:17 crc kubenswrapper[4934]: I1227 08:05:17.926557 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.360704034 podStartE2EDuration="8.926541915s" podCreationTimestamp="2025-12-27 08:05:09 +0000 UTC" firstStartedPulling="2025-12-27 08:05:10.744193396 +0000 UTC m=+1371.564633980" lastFinishedPulling="2025-12-27 08:05:17.310031267 +0000 UTC m=+1378.130471861" observedRunningTime="2025-12-27 08:05:17.924454434 +0000 UTC m=+1378.744895038" watchObservedRunningTime="2025-12-27 08:05:17.926541915 +0000 UTC m=+1378.746982509" Dec 27 08:05:18 crc kubenswrapper[4934]: I1227 08:05:18.018441 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-859445d775-g44pr"] Dec 27 08:05:18 crc kubenswrapper[4934]: W1227 08:05:18.042376 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f083ef4_9928_400b_969e_1df70b3e9475.slice/crio-9b1f168d8fe8c413ffc8c681dd7a103332622ec5d290f0ca7ca04d5e7ab6959b WatchSource:0}: Error finding container 9b1f168d8fe8c413ffc8c681dd7a103332622ec5d290f0ca7ca04d5e7ab6959b: Status 404 returned error can't find the container with id 9b1f168d8fe8c413ffc8c681dd7a103332622ec5d290f0ca7ca04d5e7ab6959b Dec 27 08:05:18 crc kubenswrapper[4934]: I1227 08:05:18.261985 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-654859b66d-km642"] Dec 27 08:05:18 crc kubenswrapper[4934]: I1227 08:05:18.272463 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-58d7546b8b-lqdtc"] Dec 27 08:05:18 crc kubenswrapper[4934]: I1227 08:05:18.282970 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-xlm2k"] Dec 27 08:05:18 crc kubenswrapper[4934]: W1227 08:05:18.290214 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod327eb20e_e905_48a6_99b8_9209813e3330.slice/crio-6cdbf81a31377bc5e87d4e32aeaa967e4c730346fc3ade778edeb7ee84f4ce3b WatchSource:0}: Error finding container 6cdbf81a31377bc5e87d4e32aeaa967e4c730346fc3ade778edeb7ee84f4ce3b: Status 404 returned error can't find the container with id 6cdbf81a31377bc5e87d4e32aeaa967e4c730346fc3ade778edeb7ee84f4ce3b Dec 27 08:05:18 crc kubenswrapper[4934]: I1227 08:05:18.912410 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-654859b66d-km642" event={"ID":"700034e0-f221-4bca-a973-a6d49c7dc406","Type":"ContainerStarted","Data":"fca1004fa8c0d46a138a4b8482fa1e25a1a874dd33970c9bb32fdb338944c363"} Dec 27 08:05:18 crc kubenswrapper[4934]: I1227 08:05:18.920717 4934 generic.go:334] "Generic (PLEG): container finished" podID="72d0ed22-31be-4048-b031-139856c584c5" containerID="dd3aa6e013a5c434c3c2c70132d259a4f0966aee7c14cdfa3dc77dd5572fc961" exitCode=0 Dec 27 08:05:18 crc kubenswrapper[4934]: I1227 08:05:18.920805 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rrj89" event={"ID":"72d0ed22-31be-4048-b031-139856c584c5","Type":"ContainerDied","Data":"dd3aa6e013a5c434c3c2c70132d259a4f0966aee7c14cdfa3dc77dd5572fc961"} Dec 27 08:05:18 crc kubenswrapper[4934]: I1227 08:05:18.926452 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58d7546b8b-lqdtc" event={"ID":"327eb20e-e905-48a6-99b8-9209813e3330","Type":"ContainerStarted","Data":"440c2ffcdd51fdd7c00a40b0c3598aa5f6961ca95292c78d0f54a9362030242c"} Dec 27 08:05:18 crc kubenswrapper[4934]: I1227 08:05:18.926502 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58d7546b8b-lqdtc" event={"ID":"327eb20e-e905-48a6-99b8-9209813e3330","Type":"ContainerStarted","Data":"f7c0182c802c5478d3e949a8b8cfad73ed9ed473d0b1b437a975df41dd930ac0"} Dec 27 08:05:18 crc kubenswrapper[4934]: I1227 08:05:18.926513 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58d7546b8b-lqdtc" event={"ID":"327eb20e-e905-48a6-99b8-9209813e3330","Type":"ContainerStarted","Data":"6cdbf81a31377bc5e87d4e32aeaa967e4c730346fc3ade778edeb7ee84f4ce3b"} Dec 27 08:05:18 crc kubenswrapper[4934]: I1227 08:05:18.927425 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:18 crc kubenswrapper[4934]: I1227 08:05:18.927455 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:18 crc kubenswrapper[4934]: I1227 08:05:18.950726 4934 generic.go:334] "Generic (PLEG): container finished" podID="48fe2613-0255-4c5c-9b4e-7b26e4360e1e" containerID="35418acf1c777ba3e05525b936fe7f474f2b081380a130fb0d0ca73582055ac1" exitCode=0 Dec 27 08:05:18 crc kubenswrapper[4934]: I1227 08:05:18.951879 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" event={"ID":"48fe2613-0255-4c5c-9b4e-7b26e4360e1e","Type":"ContainerDied","Data":"35418acf1c777ba3e05525b936fe7f474f2b081380a130fb0d0ca73582055ac1"} Dec 27 08:05:18 crc kubenswrapper[4934]: I1227 08:05:18.951913 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" event={"ID":"48fe2613-0255-4c5c-9b4e-7b26e4360e1e","Type":"ContainerStarted","Data":"eed4233ecf91f0fcf6093d89e899d371f96055d1271727389bfa409577eb3480"} Dec 27 08:05:18 crc kubenswrapper[4934]: I1227 08:05:18.959530 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-859445d775-g44pr" event={"ID":"3f083ef4-9928-400b-969e-1df70b3e9475","Type":"ContainerStarted","Data":"9b1f168d8fe8c413ffc8c681dd7a103332622ec5d290f0ca7ca04d5e7ab6959b"} Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.010429 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-58d7546b8b-lqdtc" podStartSLOduration=2.010405984 podStartE2EDuration="2.010405984s" podCreationTimestamp="2025-12-27 08:05:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:05:18.972475299 +0000 UTC m=+1379.792915903" watchObservedRunningTime="2025-12-27 08:05:19.010405984 +0000 UTC m=+1379.830846588" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.387772 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-788c8bbcbc-rtffl"] Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.389915 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.395495 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.395702 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.398299 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.411632 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-788c8bbcbc-rtffl"] Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.430613 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/930bd1af-796b-409e-8bae-d828b871edad-public-tls-certs\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.430692 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/930bd1af-796b-409e-8bae-d828b871edad-config-data\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.430713 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/930bd1af-796b-409e-8bae-d828b871edad-log-httpd\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.430769 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/930bd1af-796b-409e-8bae-d828b871edad-internal-tls-certs\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.431117 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97h8x\" (UniqueName: \"kubernetes.io/projected/930bd1af-796b-409e-8bae-d828b871edad-kube-api-access-97h8x\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.431166 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/930bd1af-796b-409e-8bae-d828b871edad-run-httpd\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.431220 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/930bd1af-796b-409e-8bae-d828b871edad-etc-swift\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.431518 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/930bd1af-796b-409e-8bae-d828b871edad-combined-ca-bundle\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.535127 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/930bd1af-796b-409e-8bae-d828b871edad-combined-ca-bundle\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.535204 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/930bd1af-796b-409e-8bae-d828b871edad-public-tls-certs\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.535311 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/930bd1af-796b-409e-8bae-d828b871edad-config-data\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.535332 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/930bd1af-796b-409e-8bae-d828b871edad-log-httpd\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.535381 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/930bd1af-796b-409e-8bae-d828b871edad-internal-tls-certs\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.535433 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97h8x\" (UniqueName: \"kubernetes.io/projected/930bd1af-796b-409e-8bae-d828b871edad-kube-api-access-97h8x\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.535450 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/930bd1af-796b-409e-8bae-d828b871edad-run-httpd\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.535469 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/930bd1af-796b-409e-8bae-d828b871edad-etc-swift\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.536411 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/930bd1af-796b-409e-8bae-d828b871edad-log-httpd\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.536514 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/930bd1af-796b-409e-8bae-d828b871edad-run-httpd\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.537353 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.537449 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.540129 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.548884 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/930bd1af-796b-409e-8bae-d828b871edad-combined-ca-bundle\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.550182 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/930bd1af-796b-409e-8bae-d828b871edad-internal-tls-certs\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.553346 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/930bd1af-796b-409e-8bae-d828b871edad-config-data\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.555774 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/930bd1af-796b-409e-8bae-d828b871edad-public-tls-certs\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.556664 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97h8x\" (UniqueName: \"kubernetes.io/projected/930bd1af-796b-409e-8bae-d828b871edad-kube-api-access-97h8x\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.557262 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/930bd1af-796b-409e-8bae-d828b871edad-etc-swift\") pod \"swift-proxy-788c8bbcbc-rtffl\" (UID: \"930bd1af-796b-409e-8bae-d828b871edad\") " pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.719934 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.987066 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" event={"ID":"48fe2613-0255-4c5c-9b4e-7b26e4360e1e","Type":"ContainerStarted","Data":"a243b21842e0c612009fa7d2d0f360600d0935896aed44c85c4b3743566d89a2"} Dec 27 08:05:19 crc kubenswrapper[4934]: I1227 08:05:19.987508 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.021670 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" podStartSLOduration=3.021650944 podStartE2EDuration="3.021650944s" podCreationTimestamp="2025-12-27 08:05:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:05:20.018075025 +0000 UTC m=+1380.838515629" watchObservedRunningTime="2025-12-27 08:05:20.021650944 +0000 UTC m=+1380.842091538" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.370569 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-788c8bbcbc-rtffl"] Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.755473 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-76d79bd746-4mrv6"] Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.759887 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.761446 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.762663 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.767551 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-76d79bd746-4mrv6"] Dec 27 08:05:20 crc kubenswrapper[4934]: W1227 08:05:20.803258 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod930bd1af_796b_409e_8bae_d828b871edad.slice/crio-43fd42a6320d49bdde5edcdc2daf274907ddece45d01c87efb9237d15b85f483 WatchSource:0}: Error finding container 43fd42a6320d49bdde5edcdc2daf274907ddece45d01c87efb9237d15b85f483: Status 404 returned error can't find the container with id 43fd42a6320d49bdde5edcdc2daf274907ddece45d01c87efb9237d15b85f483 Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.876524 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/00fe7217-c021-410c-9fca-36b69e513e92-internal-tls-certs\") pod \"barbican-api-76d79bd746-4mrv6\" (UID: \"00fe7217-c021-410c-9fca-36b69e513e92\") " pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.876569 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmwrp\" (UniqueName: \"kubernetes.io/projected/00fe7217-c021-410c-9fca-36b69e513e92-kube-api-access-nmwrp\") pod \"barbican-api-76d79bd746-4mrv6\" (UID: \"00fe7217-c021-410c-9fca-36b69e513e92\") " pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.876594 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/00fe7217-c021-410c-9fca-36b69e513e92-public-tls-certs\") pod \"barbican-api-76d79bd746-4mrv6\" (UID: \"00fe7217-c021-410c-9fca-36b69e513e92\") " pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.876674 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/00fe7217-c021-410c-9fca-36b69e513e92-config-data-custom\") pod \"barbican-api-76d79bd746-4mrv6\" (UID: \"00fe7217-c021-410c-9fca-36b69e513e92\") " pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.876699 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00fe7217-c021-410c-9fca-36b69e513e92-config-data\") pod \"barbican-api-76d79bd746-4mrv6\" (UID: \"00fe7217-c021-410c-9fca-36b69e513e92\") " pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.876728 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00fe7217-c021-410c-9fca-36b69e513e92-combined-ca-bundle\") pod \"barbican-api-76d79bd746-4mrv6\" (UID: \"00fe7217-c021-410c-9fca-36b69e513e92\") " pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.876744 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00fe7217-c021-410c-9fca-36b69e513e92-logs\") pod \"barbican-api-76d79bd746-4mrv6\" (UID: \"00fe7217-c021-410c-9fca-36b69e513e92\") " pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.928043 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rrj89" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.978798 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-config-data\") pod \"72d0ed22-31be-4048-b031-139856c584c5\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.978927 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-db-sync-config-data\") pod \"72d0ed22-31be-4048-b031-139856c584c5\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.979110 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8mxj\" (UniqueName: \"kubernetes.io/projected/72d0ed22-31be-4048-b031-139856c584c5-kube-api-access-l8mxj\") pod \"72d0ed22-31be-4048-b031-139856c584c5\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.979201 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-scripts\") pod \"72d0ed22-31be-4048-b031-139856c584c5\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.979247 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/72d0ed22-31be-4048-b031-139856c584c5-etc-machine-id\") pod \"72d0ed22-31be-4048-b031-139856c584c5\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.979284 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-combined-ca-bundle\") pod \"72d0ed22-31be-4048-b031-139856c584c5\" (UID: \"72d0ed22-31be-4048-b031-139856c584c5\") " Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.979541 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/00fe7217-c021-410c-9fca-36b69e513e92-internal-tls-certs\") pod \"barbican-api-76d79bd746-4mrv6\" (UID: \"00fe7217-c021-410c-9fca-36b69e513e92\") " pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.979575 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmwrp\" (UniqueName: \"kubernetes.io/projected/00fe7217-c021-410c-9fca-36b69e513e92-kube-api-access-nmwrp\") pod \"barbican-api-76d79bd746-4mrv6\" (UID: \"00fe7217-c021-410c-9fca-36b69e513e92\") " pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.979601 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/00fe7217-c021-410c-9fca-36b69e513e92-public-tls-certs\") pod \"barbican-api-76d79bd746-4mrv6\" (UID: \"00fe7217-c021-410c-9fca-36b69e513e92\") " pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.979732 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/00fe7217-c021-410c-9fca-36b69e513e92-config-data-custom\") pod \"barbican-api-76d79bd746-4mrv6\" (UID: \"00fe7217-c021-410c-9fca-36b69e513e92\") " pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.979762 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00fe7217-c021-410c-9fca-36b69e513e92-config-data\") pod \"barbican-api-76d79bd746-4mrv6\" (UID: \"00fe7217-c021-410c-9fca-36b69e513e92\") " pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.979805 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00fe7217-c021-410c-9fca-36b69e513e92-combined-ca-bundle\") pod \"barbican-api-76d79bd746-4mrv6\" (UID: \"00fe7217-c021-410c-9fca-36b69e513e92\") " pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.979823 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00fe7217-c021-410c-9fca-36b69e513e92-logs\") pod \"barbican-api-76d79bd746-4mrv6\" (UID: \"00fe7217-c021-410c-9fca-36b69e513e92\") " pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.980318 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00fe7217-c021-410c-9fca-36b69e513e92-logs\") pod \"barbican-api-76d79bd746-4mrv6\" (UID: \"00fe7217-c021-410c-9fca-36b69e513e92\") " pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.987141 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/72d0ed22-31be-4048-b031-139856c584c5-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "72d0ed22-31be-4048-b031-139856c584c5" (UID: "72d0ed22-31be-4048-b031-139856c584c5"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.988909 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00fe7217-c021-410c-9fca-36b69e513e92-config-data\") pod \"barbican-api-76d79bd746-4mrv6\" (UID: \"00fe7217-c021-410c-9fca-36b69e513e92\") " pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.992717 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/00fe7217-c021-410c-9fca-36b69e513e92-public-tls-certs\") pod \"barbican-api-76d79bd746-4mrv6\" (UID: \"00fe7217-c021-410c-9fca-36b69e513e92\") " pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.996005 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72d0ed22-31be-4048-b031-139856c584c5-kube-api-access-l8mxj" (OuterVolumeSpecName: "kube-api-access-l8mxj") pod "72d0ed22-31be-4048-b031-139856c584c5" (UID: "72d0ed22-31be-4048-b031-139856c584c5"). InnerVolumeSpecName "kube-api-access-l8mxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:20 crc kubenswrapper[4934]: I1227 08:05:20.996347 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "72d0ed22-31be-4048-b031-139856c584c5" (UID: "72d0ed22-31be-4048-b031-139856c584c5"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.000869 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-scripts" (OuterVolumeSpecName: "scripts") pod "72d0ed22-31be-4048-b031-139856c584c5" (UID: "72d0ed22-31be-4048-b031-139856c584c5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.007469 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/00fe7217-c021-410c-9fca-36b69e513e92-config-data-custom\") pod \"barbican-api-76d79bd746-4mrv6\" (UID: \"00fe7217-c021-410c-9fca-36b69e513e92\") " pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.007844 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/00fe7217-c021-410c-9fca-36b69e513e92-internal-tls-certs\") pod \"barbican-api-76d79bd746-4mrv6\" (UID: \"00fe7217-c021-410c-9fca-36b69e513e92\") " pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.011601 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00fe7217-c021-410c-9fca-36b69e513e92-combined-ca-bundle\") pod \"barbican-api-76d79bd746-4mrv6\" (UID: \"00fe7217-c021-410c-9fca-36b69e513e92\") " pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.011735 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-rrj89" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.012230 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-rrj89" event={"ID":"72d0ed22-31be-4048-b031-139856c584c5","Type":"ContainerDied","Data":"4fc61ed7d15598ee35bfff4f23aede2f58447a5d5682e3e1cdc3d21129cc516b"} Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.012267 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fc61ed7d15598ee35bfff4f23aede2f58447a5d5682e3e1cdc3d21129cc516b" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.014759 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-788c8bbcbc-rtffl" event={"ID":"930bd1af-796b-409e-8bae-d828b871edad","Type":"ContainerStarted","Data":"43fd42a6320d49bdde5edcdc2daf274907ddece45d01c87efb9237d15b85f483"} Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.018309 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmwrp\" (UniqueName: \"kubernetes.io/projected/00fe7217-c021-410c-9fca-36b69e513e92-kube-api-access-nmwrp\") pod \"barbican-api-76d79bd746-4mrv6\" (UID: \"00fe7217-c021-410c-9fca-36b69e513e92\") " pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.066021 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72d0ed22-31be-4048-b031-139856c584c5" (UID: "72d0ed22-31be-4048-b031-139856c584c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.077008 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.083616 4934 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.083644 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8mxj\" (UniqueName: \"kubernetes.io/projected/72d0ed22-31be-4048-b031-139856c584c5-kube-api-access-l8mxj\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.083655 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.083665 4934 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/72d0ed22-31be-4048-b031-139856c584c5-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.083674 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.120785 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-config-data" (OuterVolumeSpecName: "config-data") pod "72d0ed22-31be-4048-b031-139856c584c5" (UID: "72d0ed22-31be-4048-b031-139856c584c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.188153 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72d0ed22-31be-4048-b031-139856c584c5-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.232131 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 27 08:05:21 crc kubenswrapper[4934]: E1227 08:05:21.232778 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72d0ed22-31be-4048-b031-139856c584c5" containerName="cinder-db-sync" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.232796 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="72d0ed22-31be-4048-b031-139856c584c5" containerName="cinder-db-sync" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.233110 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="72d0ed22-31be-4048-b031-139856c584c5" containerName="cinder-db-sync" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.234660 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.246469 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-xlm2k"] Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.259348 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.264269 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.290756 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.291226 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whj85\" (UniqueName: \"kubernetes.io/projected/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-kube-api-access-whj85\") pod \"cinder-scheduler-0\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.291275 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-scripts\") pod \"cinder-scheduler-0\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.291507 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.291619 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.291693 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-config-data\") pod \"cinder-scheduler-0\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.372152 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-5gt6v"] Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.373969 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.394290 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.394391 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whj85\" (UniqueName: \"kubernetes.io/projected/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-kube-api-access-whj85\") pod \"cinder-scheduler-0\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.394412 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-scripts\") pod \"cinder-scheduler-0\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.394497 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.394545 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.394581 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-config-data\") pod \"cinder-scheduler-0\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.399834 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.407661 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.419155 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-config-data\") pod \"cinder-scheduler-0\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.434864 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whj85\" (UniqueName: \"kubernetes.io/projected/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-kube-api-access-whj85\") pod \"cinder-scheduler-0\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.434929 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-5gt6v"] Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.439571 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-scripts\") pod \"cinder-scheduler-0\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.458729 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.546938 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-5gt6v\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.547012 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-5gt6v\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.547062 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-5gt6v\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.547112 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-dns-svc\") pod \"dnsmasq-dns-6578955fd5-5gt6v\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.547162 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-config\") pod \"dnsmasq-dns-6578955fd5-5gt6v\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.547280 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92tfz\" (UniqueName: \"kubernetes.io/projected/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-kube-api-access-92tfz\") pod \"dnsmasq-dns-6578955fd5-5gt6v\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.566580 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.646830 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.650203 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-5gt6v\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.650274 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-dns-svc\") pod \"dnsmasq-dns-6578955fd5-5gt6v\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.650329 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-config\") pod \"dnsmasq-dns-6578955fd5-5gt6v\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.650576 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92tfz\" (UniqueName: \"kubernetes.io/projected/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-kube-api-access-92tfz\") pod \"dnsmasq-dns-6578955fd5-5gt6v\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.650602 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-5gt6v\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.650639 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-5gt6v\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.651570 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-5gt6v\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.652041 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-5gt6v\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.653005 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-config\") pod \"dnsmasq-dns-6578955fd5-5gt6v\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.653216 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-dns-svc\") pod \"dnsmasq-dns-6578955fd5-5gt6v\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.653892 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-5gt6v\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.658459 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.662477 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.662805 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.674028 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92tfz\" (UniqueName: \"kubernetes.io/projected/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-kube-api-access-92tfz\") pod \"dnsmasq-dns-6578955fd5-5gt6v\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.708590 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.752747 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/34c0fb07-e3e3-45f0-b785-5c981a771ceb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.752812 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.752892 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdhlk\" (UniqueName: \"kubernetes.io/projected/34c0fb07-e3e3-45f0-b785-5c981a771ceb-kube-api-access-hdhlk\") pod \"cinder-api-0\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.752923 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34c0fb07-e3e3-45f0-b785-5c981a771ceb-logs\") pod \"cinder-api-0\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.753026 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-config-data\") pod \"cinder-api-0\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.753099 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-config-data-custom\") pod \"cinder-api-0\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.753116 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-scripts\") pod \"cinder-api-0\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.855230 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-config-data-custom\") pod \"cinder-api-0\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.855272 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-scripts\") pod \"cinder-api-0\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.855366 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/34c0fb07-e3e3-45f0-b785-5c981a771ceb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.855416 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.855463 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdhlk\" (UniqueName: \"kubernetes.io/projected/34c0fb07-e3e3-45f0-b785-5c981a771ceb-kube-api-access-hdhlk\") pod \"cinder-api-0\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.855494 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34c0fb07-e3e3-45f0-b785-5c981a771ceb-logs\") pod \"cinder-api-0\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.855633 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-config-data\") pod \"cinder-api-0\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.856353 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/34c0fb07-e3e3-45f0-b785-5c981a771ceb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.856412 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34c0fb07-e3e3-45f0-b785-5c981a771ceb-logs\") pod \"cinder-api-0\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.860139 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-scripts\") pod \"cinder-api-0\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.860640 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-config-data-custom\") pod \"cinder-api-0\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.862668 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-config-data\") pod \"cinder-api-0\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.871807 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.873828 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdhlk\" (UniqueName: \"kubernetes.io/projected/34c0fb07-e3e3-45f0-b785-5c981a771ceb-kube-api-access-hdhlk\") pod \"cinder-api-0\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " pod="openstack/cinder-api-0" Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.963004 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.963503 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cef13c6c-b21d-45f8-83be-625000146643" containerName="ceilometer-central-agent" containerID="cri-o://ea63ad9ef292b297cb30f1be32c7e34a88199aa88c4d09e99e027b160549fcb0" gracePeriod=30 Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.963646 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cef13c6c-b21d-45f8-83be-625000146643" containerName="proxy-httpd" containerID="cri-o://75b87c79a50fe702df30afcea7ea1735e5bab70d914b25d800a4bf0503db1a01" gracePeriod=30 Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.963697 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cef13c6c-b21d-45f8-83be-625000146643" containerName="sg-core" containerID="cri-o://0e4457b0013fbecad3139a089d129032412a51cceb7d10b391d48c65f4f84f4f" gracePeriod=30 Dec 27 08:05:21 crc kubenswrapper[4934]: I1227 08:05:21.963728 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="cef13c6c-b21d-45f8-83be-625000146643" containerName="ceilometer-notification-agent" containerID="cri-o://157284b0f56ac0f09ace1cb65b4cc1ae55750e564536da9dd030cd2fed0a61da" gracePeriod=30 Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.020905 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.025031 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" podUID="48fe2613-0255-4c5c-9b4e-7b26e4360e1e" containerName="dnsmasq-dns" containerID="cri-o://a243b21842e0c612009fa7d2d0f360600d0935896aed44c85c4b3743566d89a2" gracePeriod=10 Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.154354 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-6bdc8c9885-7z5pz"] Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.155944 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6bdc8c9885-7z5pz" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.159416 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.159613 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.164402 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-7lm5l" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.185341 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6bdc8c9885-7z5pz"] Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.272301 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c8feac2-c2ee-484c-baed-052bd7c0b468-config-data\") pod \"heat-engine-6bdc8c9885-7z5pz\" (UID: \"2c8feac2-c2ee-484c-baed-052bd7c0b468\") " pod="openstack/heat-engine-6bdc8c9885-7z5pz" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.272345 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctnlb\" (UniqueName: \"kubernetes.io/projected/2c8feac2-c2ee-484c-baed-052bd7c0b468-kube-api-access-ctnlb\") pod \"heat-engine-6bdc8c9885-7z5pz\" (UID: \"2c8feac2-c2ee-484c-baed-052bd7c0b468\") " pod="openstack/heat-engine-6bdc8c9885-7z5pz" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.272395 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2c8feac2-c2ee-484c-baed-052bd7c0b468-config-data-custom\") pod \"heat-engine-6bdc8c9885-7z5pz\" (UID: \"2c8feac2-c2ee-484c-baed-052bd7c0b468\") " pod="openstack/heat-engine-6bdc8c9885-7z5pz" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.272561 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c8feac2-c2ee-484c-baed-052bd7c0b468-combined-ca-bundle\") pod \"heat-engine-6bdc8c9885-7z5pz\" (UID: \"2c8feac2-c2ee-484c-baed-052bd7c0b468\") " pod="openstack/heat-engine-6bdc8c9885-7z5pz" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.284290 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-f7fc9c49c-vpz65"] Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.294541 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-f7fc9c49c-vpz65" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.304729 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.316185 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-f7fc9c49c-vpz65"] Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.365780 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-5gt6v"] Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.375073 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c8feac2-c2ee-484c-baed-052bd7c0b468-combined-ca-bundle\") pod \"heat-engine-6bdc8c9885-7z5pz\" (UID: \"2c8feac2-c2ee-484c-baed-052bd7c0b468\") " pod="openstack/heat-engine-6bdc8c9885-7z5pz" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.375351 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c8feac2-c2ee-484c-baed-052bd7c0b468-config-data\") pod \"heat-engine-6bdc8c9885-7z5pz\" (UID: \"2c8feac2-c2ee-484c-baed-052bd7c0b468\") " pod="openstack/heat-engine-6bdc8c9885-7z5pz" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.375449 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctnlb\" (UniqueName: \"kubernetes.io/projected/2c8feac2-c2ee-484c-baed-052bd7c0b468-kube-api-access-ctnlb\") pod \"heat-engine-6bdc8c9885-7z5pz\" (UID: \"2c8feac2-c2ee-484c-baed-052bd7c0b468\") " pod="openstack/heat-engine-6bdc8c9885-7z5pz" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.375594 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2c8feac2-c2ee-484c-baed-052bd7c0b468-config-data-custom\") pod \"heat-engine-6bdc8c9885-7z5pz\" (UID: \"2c8feac2-c2ee-484c-baed-052bd7c0b468\") " pod="openstack/heat-engine-6bdc8c9885-7z5pz" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.381802 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c8feac2-c2ee-484c-baed-052bd7c0b468-combined-ca-bundle\") pod \"heat-engine-6bdc8c9885-7z5pz\" (UID: \"2c8feac2-c2ee-484c-baed-052bd7c0b468\") " pod="openstack/heat-engine-6bdc8c9885-7z5pz" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.385215 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c8feac2-c2ee-484c-baed-052bd7c0b468-config-data\") pod \"heat-engine-6bdc8c9885-7z5pz\" (UID: \"2c8feac2-c2ee-484c-baed-052bd7c0b468\") " pod="openstack/heat-engine-6bdc8c9885-7z5pz" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.419794 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctnlb\" (UniqueName: \"kubernetes.io/projected/2c8feac2-c2ee-484c-baed-052bd7c0b468-kube-api-access-ctnlb\") pod \"heat-engine-6bdc8c9885-7z5pz\" (UID: \"2c8feac2-c2ee-484c-baed-052bd7c0b468\") " pod="openstack/heat-engine-6bdc8c9885-7z5pz" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.420093 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2c8feac2-c2ee-484c-baed-052bd7c0b468-config-data-custom\") pod \"heat-engine-6bdc8c9885-7z5pz\" (UID: \"2c8feac2-c2ee-484c-baed-052bd7c0b468\") " pod="openstack/heat-engine-6bdc8c9885-7z5pz" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.429233 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6bdc8c9885-7z5pz" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.435480 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5cc8cdc7b7-mmj42"] Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.436882 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5cc8cdc7b7-mmj42" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.441533 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.478463 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjc2p\" (UniqueName: \"kubernetes.io/projected/7b983753-0d99-48aa-82e9-c5a0643cf4c3-kube-api-access-wjc2p\") pod \"heat-cfnapi-f7fc9c49c-vpz65\" (UID: \"7b983753-0d99-48aa-82e9-c5a0643cf4c3\") " pod="openstack/heat-cfnapi-f7fc9c49c-vpz65" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.478644 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b983753-0d99-48aa-82e9-c5a0643cf4c3-config-data-custom\") pod \"heat-cfnapi-f7fc9c49c-vpz65\" (UID: \"7b983753-0d99-48aa-82e9-c5a0643cf4c3\") " pod="openstack/heat-cfnapi-f7fc9c49c-vpz65" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.478687 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b983753-0d99-48aa-82e9-c5a0643cf4c3-combined-ca-bundle\") pod \"heat-cfnapi-f7fc9c49c-vpz65\" (UID: \"7b983753-0d99-48aa-82e9-c5a0643cf4c3\") " pod="openstack/heat-cfnapi-f7fc9c49c-vpz65" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.478791 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b983753-0d99-48aa-82e9-c5a0643cf4c3-config-data\") pod \"heat-cfnapi-f7fc9c49c-vpz65\" (UID: \"7b983753-0d99-48aa-82e9-c5a0643cf4c3\") " pod="openstack/heat-cfnapi-f7fc9c49c-vpz65" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.500070 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5cc8cdc7b7-mmj42"] Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.520668 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-l22fd"] Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.529697 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.557911 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-l22fd"] Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.584713 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b983753-0d99-48aa-82e9-c5a0643cf4c3-combined-ca-bundle\") pod \"heat-cfnapi-f7fc9c49c-vpz65\" (UID: \"7b983753-0d99-48aa-82e9-c5a0643cf4c3\") " pod="openstack/heat-cfnapi-f7fc9c49c-vpz65" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.584892 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a23b3091-07c0-4617-9fb9-de876bed5e12-combined-ca-bundle\") pod \"heat-api-5cc8cdc7b7-mmj42\" (UID: \"a23b3091-07c0-4617-9fb9-de876bed5e12\") " pod="openstack/heat-api-5cc8cdc7b7-mmj42" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.587960 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b983753-0d99-48aa-82e9-c5a0643cf4c3-config-data\") pod \"heat-cfnapi-f7fc9c49c-vpz65\" (UID: \"7b983753-0d99-48aa-82e9-c5a0643cf4c3\") " pod="openstack/heat-cfnapi-f7fc9c49c-vpz65" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.587989 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a23b3091-07c0-4617-9fb9-de876bed5e12-config-data\") pod \"heat-api-5cc8cdc7b7-mmj42\" (UID: \"a23b3091-07c0-4617-9fb9-de876bed5e12\") " pod="openstack/heat-api-5cc8cdc7b7-mmj42" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.588098 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjc2p\" (UniqueName: \"kubernetes.io/projected/7b983753-0d99-48aa-82e9-c5a0643cf4c3-kube-api-access-wjc2p\") pod \"heat-cfnapi-f7fc9c49c-vpz65\" (UID: \"7b983753-0d99-48aa-82e9-c5a0643cf4c3\") " pod="openstack/heat-cfnapi-f7fc9c49c-vpz65" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.588157 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdxpj\" (UniqueName: \"kubernetes.io/projected/a23b3091-07c0-4617-9fb9-de876bed5e12-kube-api-access-pdxpj\") pod \"heat-api-5cc8cdc7b7-mmj42\" (UID: \"a23b3091-07c0-4617-9fb9-de876bed5e12\") " pod="openstack/heat-api-5cc8cdc7b7-mmj42" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.589108 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a23b3091-07c0-4617-9fb9-de876bed5e12-config-data-custom\") pod \"heat-api-5cc8cdc7b7-mmj42\" (UID: \"a23b3091-07c0-4617-9fb9-de876bed5e12\") " pod="openstack/heat-api-5cc8cdc7b7-mmj42" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.589165 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b983753-0d99-48aa-82e9-c5a0643cf4c3-config-data-custom\") pod \"heat-cfnapi-f7fc9c49c-vpz65\" (UID: \"7b983753-0d99-48aa-82e9-c5a0643cf4c3\") " pod="openstack/heat-cfnapi-f7fc9c49c-vpz65" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.592379 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b983753-0d99-48aa-82e9-c5a0643cf4c3-config-data\") pod \"heat-cfnapi-f7fc9c49c-vpz65\" (UID: \"7b983753-0d99-48aa-82e9-c5a0643cf4c3\") " pod="openstack/heat-cfnapi-f7fc9c49c-vpz65" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.609301 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b983753-0d99-48aa-82e9-c5a0643cf4c3-config-data-custom\") pod \"heat-cfnapi-f7fc9c49c-vpz65\" (UID: \"7b983753-0d99-48aa-82e9-c5a0643cf4c3\") " pod="openstack/heat-cfnapi-f7fc9c49c-vpz65" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.613831 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjc2p\" (UniqueName: \"kubernetes.io/projected/7b983753-0d99-48aa-82e9-c5a0643cf4c3-kube-api-access-wjc2p\") pod \"heat-cfnapi-f7fc9c49c-vpz65\" (UID: \"7b983753-0d99-48aa-82e9-c5a0643cf4c3\") " pod="openstack/heat-cfnapi-f7fc9c49c-vpz65" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.624170 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b983753-0d99-48aa-82e9-c5a0643cf4c3-combined-ca-bundle\") pod \"heat-cfnapi-f7fc9c49c-vpz65\" (UID: \"7b983753-0d99-48aa-82e9-c5a0643cf4c3\") " pod="openstack/heat-cfnapi-f7fc9c49c-vpz65" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.701685 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-ovsdbserver-sb\") pod \"dnsmasq-dns-688b9f5b49-l22fd\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.701777 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a23b3091-07c0-4617-9fb9-de876bed5e12-combined-ca-bundle\") pod \"heat-api-5cc8cdc7b7-mmj42\" (UID: \"a23b3091-07c0-4617-9fb9-de876bed5e12\") " pod="openstack/heat-api-5cc8cdc7b7-mmj42" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.701806 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a23b3091-07c0-4617-9fb9-de876bed5e12-config-data\") pod \"heat-api-5cc8cdc7b7-mmj42\" (UID: \"a23b3091-07c0-4617-9fb9-de876bed5e12\") " pod="openstack/heat-api-5cc8cdc7b7-mmj42" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.701826 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-dns-svc\") pod \"dnsmasq-dns-688b9f5b49-l22fd\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.701851 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-config\") pod \"dnsmasq-dns-688b9f5b49-l22fd\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.701871 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpk9k\" (UniqueName: \"kubernetes.io/projected/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-kube-api-access-mpk9k\") pod \"dnsmasq-dns-688b9f5b49-l22fd\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.701912 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdxpj\" (UniqueName: \"kubernetes.io/projected/a23b3091-07c0-4617-9fb9-de876bed5e12-kube-api-access-pdxpj\") pod \"heat-api-5cc8cdc7b7-mmj42\" (UID: \"a23b3091-07c0-4617-9fb9-de876bed5e12\") " pod="openstack/heat-api-5cc8cdc7b7-mmj42" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.701993 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-ovsdbserver-nb\") pod \"dnsmasq-dns-688b9f5b49-l22fd\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.702023 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-dns-swift-storage-0\") pod \"dnsmasq-dns-688b9f5b49-l22fd\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.702045 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a23b3091-07c0-4617-9fb9-de876bed5e12-config-data-custom\") pod \"heat-api-5cc8cdc7b7-mmj42\" (UID: \"a23b3091-07c0-4617-9fb9-de876bed5e12\") " pod="openstack/heat-api-5cc8cdc7b7-mmj42" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.715959 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a23b3091-07c0-4617-9fb9-de876bed5e12-combined-ca-bundle\") pod \"heat-api-5cc8cdc7b7-mmj42\" (UID: \"a23b3091-07c0-4617-9fb9-de876bed5e12\") " pod="openstack/heat-api-5cc8cdc7b7-mmj42" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.716880 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a23b3091-07c0-4617-9fb9-de876bed5e12-config-data\") pod \"heat-api-5cc8cdc7b7-mmj42\" (UID: \"a23b3091-07c0-4617-9fb9-de876bed5e12\") " pod="openstack/heat-api-5cc8cdc7b7-mmj42" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.732920 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a23b3091-07c0-4617-9fb9-de876bed5e12-config-data-custom\") pod \"heat-api-5cc8cdc7b7-mmj42\" (UID: \"a23b3091-07c0-4617-9fb9-de876bed5e12\") " pod="openstack/heat-api-5cc8cdc7b7-mmj42" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.753501 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-f7fc9c49c-vpz65" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.756216 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdxpj\" (UniqueName: \"kubernetes.io/projected/a23b3091-07c0-4617-9fb9-de876bed5e12-kube-api-access-pdxpj\") pod \"heat-api-5cc8cdc7b7-mmj42\" (UID: \"a23b3091-07c0-4617-9fb9-de876bed5e12\") " pod="openstack/heat-api-5cc8cdc7b7-mmj42" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.792439 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5cc8cdc7b7-mmj42" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.804843 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-dns-svc\") pod \"dnsmasq-dns-688b9f5b49-l22fd\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.804897 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-config\") pod \"dnsmasq-dns-688b9f5b49-l22fd\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.804923 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpk9k\" (UniqueName: \"kubernetes.io/projected/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-kube-api-access-mpk9k\") pod \"dnsmasq-dns-688b9f5b49-l22fd\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.805063 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-ovsdbserver-nb\") pod \"dnsmasq-dns-688b9f5b49-l22fd\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.805704 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-dns-swift-storage-0\") pod \"dnsmasq-dns-688b9f5b49-l22fd\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.806206 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-ovsdbserver-sb\") pod \"dnsmasq-dns-688b9f5b49-l22fd\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.806828 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-dns-swift-storage-0\") pod \"dnsmasq-dns-688b9f5b49-l22fd\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.806016 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-dns-svc\") pod \"dnsmasq-dns-688b9f5b49-l22fd\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.806977 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-ovsdbserver-nb\") pod \"dnsmasq-dns-688b9f5b49-l22fd\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.807003 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-config\") pod \"dnsmasq-dns-688b9f5b49-l22fd\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.807530 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-ovsdbserver-sb\") pod \"dnsmasq-dns-688b9f5b49-l22fd\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.843136 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpk9k\" (UniqueName: \"kubernetes.io/projected/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-kube-api-access-mpk9k\") pod \"dnsmasq-dns-688b9f5b49-l22fd\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:22 crc kubenswrapper[4934]: I1227 08:05:22.868368 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.120359 4934 generic.go:334] "Generic (PLEG): container finished" podID="48fe2613-0255-4c5c-9b4e-7b26e4360e1e" containerID="a243b21842e0c612009fa7d2d0f360600d0935896aed44c85c4b3743566d89a2" exitCode=0 Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.120700 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" event={"ID":"48fe2613-0255-4c5c-9b4e-7b26e4360e1e","Type":"ContainerDied","Data":"a243b21842e0c612009fa7d2d0f360600d0935896aed44c85c4b3743566d89a2"} Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.152420 4934 generic.go:334] "Generic (PLEG): container finished" podID="cef13c6c-b21d-45f8-83be-625000146643" containerID="75b87c79a50fe702df30afcea7ea1735e5bab70d914b25d800a4bf0503db1a01" exitCode=0 Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.152451 4934 generic.go:334] "Generic (PLEG): container finished" podID="cef13c6c-b21d-45f8-83be-625000146643" containerID="0e4457b0013fbecad3139a089d129032412a51cceb7d10b391d48c65f4f84f4f" exitCode=2 Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.152460 4934 generic.go:334] "Generic (PLEG): container finished" podID="cef13c6c-b21d-45f8-83be-625000146643" containerID="157284b0f56ac0f09ace1cb65b4cc1ae55750e564536da9dd030cd2fed0a61da" exitCode=0 Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.152467 4934 generic.go:334] "Generic (PLEG): container finished" podID="cef13c6c-b21d-45f8-83be-625000146643" containerID="ea63ad9ef292b297cb30f1be32c7e34a88199aa88c4d09e99e027b160549fcb0" exitCode=0 Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.152488 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cef13c6c-b21d-45f8-83be-625000146643","Type":"ContainerDied","Data":"75b87c79a50fe702df30afcea7ea1735e5bab70d914b25d800a4bf0503db1a01"} Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.152517 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cef13c6c-b21d-45f8-83be-625000146643","Type":"ContainerDied","Data":"0e4457b0013fbecad3139a089d129032412a51cceb7d10b391d48c65f4f84f4f"} Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.152528 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cef13c6c-b21d-45f8-83be-625000146643","Type":"ContainerDied","Data":"157284b0f56ac0f09ace1cb65b4cc1ae55750e564536da9dd030cd2fed0a61da"} Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.152538 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cef13c6c-b21d-45f8-83be-625000146643","Type":"ContainerDied","Data":"ea63ad9ef292b297cb30f1be32c7e34a88199aa88c4d09e99e027b160549fcb0"} Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.426007 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.440524 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-config\") pod \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.440618 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-dns-svc\") pod \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.440668 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-ovsdbserver-nb\") pod \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.440823 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-ovsdbserver-sb\") pod \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.440913 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdjlf\" (UniqueName: \"kubernetes.io/projected/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-kube-api-access-fdjlf\") pod \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.441059 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-dns-swift-storage-0\") pod \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\" (UID: \"48fe2613-0255-4c5c-9b4e-7b26e4360e1e\") " Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.501889 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-kube-api-access-fdjlf" (OuterVolumeSpecName: "kube-api-access-fdjlf") pod "48fe2613-0255-4c5c-9b4e-7b26e4360e1e" (UID: "48fe2613-0255-4c5c-9b4e-7b26e4360e1e"). InnerVolumeSpecName "kube-api-access-fdjlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.549827 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdjlf\" (UniqueName: \"kubernetes.io/projected/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-kube-api-access-fdjlf\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.615401 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-76d79bd746-4mrv6"] Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.655700 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-5gt6v"] Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.798101 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.871588 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cef13c6c-b21d-45f8-83be-625000146643-run-httpd\") pod \"cef13c6c-b21d-45f8-83be-625000146643\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.871630 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cef13c6c-b21d-45f8-83be-625000146643-log-httpd\") pod \"cef13c6c-b21d-45f8-83be-625000146643\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.871689 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-sg-core-conf-yaml\") pod \"cef13c6c-b21d-45f8-83be-625000146643\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.871739 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-combined-ca-bundle\") pod \"cef13c6c-b21d-45f8-83be-625000146643\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.871775 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-scripts\") pod \"cef13c6c-b21d-45f8-83be-625000146643\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.875897 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-config-data\") pod \"cef13c6c-b21d-45f8-83be-625000146643\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.876319 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85ssg\" (UniqueName: \"kubernetes.io/projected/cef13c6c-b21d-45f8-83be-625000146643-kube-api-access-85ssg\") pod \"cef13c6c-b21d-45f8-83be-625000146643\" (UID: \"cef13c6c-b21d-45f8-83be-625000146643\") " Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.877686 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cef13c6c-b21d-45f8-83be-625000146643-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "cef13c6c-b21d-45f8-83be-625000146643" (UID: "cef13c6c-b21d-45f8-83be-625000146643"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.884985 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cef13c6c-b21d-45f8-83be-625000146643-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "cef13c6c-b21d-45f8-83be-625000146643" (UID: "cef13c6c-b21d-45f8-83be-625000146643"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.893470 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-scripts" (OuterVolumeSpecName: "scripts") pod "cef13c6c-b21d-45f8-83be-625000146643" (UID: "cef13c6c-b21d-45f8-83be-625000146643"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:23 crc kubenswrapper[4934]: I1227 08:05:23.992869 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.002512 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-f7fc9c49c-vpz65"] Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.003934 4934 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cef13c6c-b21d-45f8-83be-625000146643-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.003969 4934 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cef13c6c-b21d-45f8-83be-625000146643-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.003981 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.036432 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cef13c6c-b21d-45f8-83be-625000146643-kube-api-access-85ssg" (OuterVolumeSpecName: "kube-api-access-85ssg") pod "cef13c6c-b21d-45f8-83be-625000146643" (UID: "cef13c6c-b21d-45f8-83be-625000146643"). InnerVolumeSpecName "kube-api-access-85ssg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.051419 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.087351 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.102936 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "48fe2613-0255-4c5c-9b4e-7b26e4360e1e" (UID: "48fe2613-0255-4c5c-9b4e-7b26e4360e1e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.103053 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6bdc8c9885-7z5pz"] Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.114399 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.114430 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85ssg\" (UniqueName: \"kubernetes.io/projected/cef13c6c-b21d-45f8-83be-625000146643-kube-api-access-85ssg\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:24 crc kubenswrapper[4934]: W1227 08:05:24.124203 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda3c3edbd_c0b4_4ea6_9e3a_0350f43ced2c.slice/crio-d8722f13d72ecfef9fc81ab8b5bdbd203a46476879bb3584058882ea7edae274 WatchSource:0}: Error finding container d8722f13d72ecfef9fc81ab8b5bdbd203a46476879bb3584058882ea7edae274: Status 404 returned error can't find the container with id d8722f13d72ecfef9fc81ab8b5bdbd203a46476879bb3584058882ea7edae274 Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.164991 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "48fe2613-0255-4c5c-9b4e-7b26e4360e1e" (UID: "48fe2613-0255-4c5c-9b4e-7b26e4360e1e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.175711 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-f7fc9c49c-vpz65" event={"ID":"7b983753-0d99-48aa-82e9-c5a0643cf4c3","Type":"ContainerStarted","Data":"07df07767df3d7d1ce37f9daa6340525c65a8f871cc7713828cf4aeee4432498"} Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.188113 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" event={"ID":"48fe2613-0255-4c5c-9b4e-7b26e4360e1e","Type":"ContainerDied","Data":"eed4233ecf91f0fcf6093d89e899d371f96055d1271727389bfa409577eb3480"} Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.188173 4934 scope.go:117] "RemoveContainer" containerID="a243b21842e0c612009fa7d2d0f360600d0935896aed44c85c4b3743566d89a2" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.188477 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-xlm2k" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.212558 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-788c8bbcbc-rtffl" event={"ID":"930bd1af-796b-409e-8bae-d828b871edad","Type":"ContainerStarted","Data":"ac67778c9aca98fcaf95ee8670cdb9320ae1a2d9752951f7a187186ed7a171fc"} Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.212604 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.212613 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-788c8bbcbc-rtffl" event={"ID":"930bd1af-796b-409e-8bae-d828b871edad","Type":"ContainerStarted","Data":"586d55b6a558641c3f4d15006292bad053bcb2ad64149681ae12f193137402b1"} Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.212635 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.217428 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-config" (OuterVolumeSpecName: "config") pod "48fe2613-0255-4c5c-9b4e-7b26e4360e1e" (UID: "48fe2613-0255-4c5c-9b4e-7b26e4360e1e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.217687 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.217703 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.225020 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "cef13c6c-b21d-45f8-83be-625000146643" (UID: "cef13c6c-b21d-45f8-83be-625000146643"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.251723 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.255196 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-l22fd"] Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.255237 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"cef13c6c-b21d-45f8-83be-625000146643","Type":"ContainerDied","Data":"3ed37090c304cdcaebe47b77760b5dcdd36d526383ff2393e1885a17c60944b6"} Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.287915 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-654859b66d-km642" event={"ID":"700034e0-f221-4bca-a973-a6d49c7dc406","Type":"ContainerStarted","Data":"0ddf03915e1c47e4401b9e02f6f54c7094db1090bc21408657b4282cf0e231aa"} Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.296061 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"34c0fb07-e3e3-45f0-b785-5c981a771ceb","Type":"ContainerStarted","Data":"d8970be25c26fa6cdfb0d68853daf1b63993f7f12ebadb265a2136b3771992f4"} Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.303278 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "48fe2613-0255-4c5c-9b4e-7b26e4360e1e" (UID: "48fe2613-0255-4c5c-9b4e-7b26e4360e1e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.306131 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-859445d775-g44pr" event={"ID":"3f083ef4-9928-400b-969e-1df70b3e9475","Type":"ContainerStarted","Data":"566d98589b2bbaa6d31de7175fae5fb3d81f8fc52003db88b85196505f148d9a"} Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.317404 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c","Type":"ContainerStarted","Data":"d8722f13d72ecfef9fc81ab8b5bdbd203a46476879bb3584058882ea7edae274"} Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.321860 4934 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.321890 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.329301 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" event={"ID":"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4","Type":"ContainerStarted","Data":"0ccb34a6fc7e56df39a724b71be390590ebe6c8f5b2f399183e44e9942596ba1"} Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.334324 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5cc8cdc7b7-mmj42"] Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.338039 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-76d79bd746-4mrv6" event={"ID":"00fe7217-c021-410c-9fca-36b69e513e92","Type":"ContainerStarted","Data":"3caa41abccf104ae1c942436fc330adb7abfc6471d5b579e38ac05faef64a1af"} Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.337117 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-788c8bbcbc-rtffl" podStartSLOduration=5.337096418 podStartE2EDuration="5.337096418s" podCreationTimestamp="2025-12-27 08:05:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:05:24.259614168 +0000 UTC m=+1385.080054752" watchObservedRunningTime="2025-12-27 08:05:24.337096418 +0000 UTC m=+1385.157537012" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.342661 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "48fe2613-0255-4c5c-9b4e-7b26e4360e1e" (UID: "48fe2613-0255-4c5c-9b4e-7b26e4360e1e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.375919 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cef13c6c-b21d-45f8-83be-625000146643" (UID: "cef13c6c-b21d-45f8-83be-625000146643"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.426137 4934 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/48fe2613-0255-4c5c-9b4e-7b26e4360e1e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.426175 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.449727 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-config-data" (OuterVolumeSpecName: "config-data") pod "cef13c6c-b21d-45f8-83be-625000146643" (UID: "cef13c6c-b21d-45f8-83be-625000146643"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.528327 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cef13c6c-b21d-45f8-83be-625000146643-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.535758 4934 scope.go:117] "RemoveContainer" containerID="35418acf1c777ba3e05525b936fe7f474f2b081380a130fb0d0ca73582055ac1" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.573603 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-xlm2k"] Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.594836 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-xlm2k"] Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.609014 4934 scope.go:117] "RemoveContainer" containerID="75b87c79a50fe702df30afcea7ea1735e5bab70d914b25d800a4bf0503db1a01" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.646336 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.686583 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.708937 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:05:24 crc kubenswrapper[4934]: E1227 08:05:24.710900 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cef13c6c-b21d-45f8-83be-625000146643" containerName="ceilometer-central-agent" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.710926 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="cef13c6c-b21d-45f8-83be-625000146643" containerName="ceilometer-central-agent" Dec 27 08:05:24 crc kubenswrapper[4934]: E1227 08:05:24.710948 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cef13c6c-b21d-45f8-83be-625000146643" containerName="sg-core" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.710955 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="cef13c6c-b21d-45f8-83be-625000146643" containerName="sg-core" Dec 27 08:05:24 crc kubenswrapper[4934]: E1227 08:05:24.710972 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cef13c6c-b21d-45f8-83be-625000146643" containerName="proxy-httpd" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.710977 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="cef13c6c-b21d-45f8-83be-625000146643" containerName="proxy-httpd" Dec 27 08:05:24 crc kubenswrapper[4934]: E1227 08:05:24.710997 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48fe2613-0255-4c5c-9b4e-7b26e4360e1e" containerName="dnsmasq-dns" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.711003 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="48fe2613-0255-4c5c-9b4e-7b26e4360e1e" containerName="dnsmasq-dns" Dec 27 08:05:24 crc kubenswrapper[4934]: E1227 08:05:24.711013 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cef13c6c-b21d-45f8-83be-625000146643" containerName="ceilometer-notification-agent" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.711019 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="cef13c6c-b21d-45f8-83be-625000146643" containerName="ceilometer-notification-agent" Dec 27 08:05:24 crc kubenswrapper[4934]: E1227 08:05:24.711038 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48fe2613-0255-4c5c-9b4e-7b26e4360e1e" containerName="init" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.711045 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="48fe2613-0255-4c5c-9b4e-7b26e4360e1e" containerName="init" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.711309 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="cef13c6c-b21d-45f8-83be-625000146643" containerName="ceilometer-central-agent" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.711331 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="cef13c6c-b21d-45f8-83be-625000146643" containerName="sg-core" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.711350 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="cef13c6c-b21d-45f8-83be-625000146643" containerName="ceilometer-notification-agent" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.711406 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="cef13c6c-b21d-45f8-83be-625000146643" containerName="proxy-httpd" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.711422 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="48fe2613-0255-4c5c-9b4e-7b26e4360e1e" containerName="dnsmasq-dns" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.713854 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.718545 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.718867 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.725222 4934 scope.go:117] "RemoveContainer" containerID="0e4457b0013fbecad3139a089d129032412a51cceb7d10b391d48c65f4f84f4f" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.725595 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.783532 4934 scope.go:117] "RemoveContainer" containerID="157284b0f56ac0f09ace1cb65b4cc1ae55750e564536da9dd030cd2fed0a61da" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.808254 4934 scope.go:117] "RemoveContainer" containerID="ea63ad9ef292b297cb30f1be32c7e34a88199aa88c4d09e99e027b160549fcb0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.836216 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-scripts\") pod \"ceilometer-0\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.836271 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.836320 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.836398 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deefee78-9e75-44e2-b72a-08aeb8b97ac4-run-httpd\") pod \"ceilometer-0\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.836432 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deefee78-9e75-44e2-b72a-08aeb8b97ac4-log-httpd\") pod \"ceilometer-0\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.836468 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55fc2\" (UniqueName: \"kubernetes.io/projected/deefee78-9e75-44e2-b72a-08aeb8b97ac4-kube-api-access-55fc2\") pod \"ceilometer-0\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.836543 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-config-data\") pod \"ceilometer-0\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.939827 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.940279 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deefee78-9e75-44e2-b72a-08aeb8b97ac4-run-httpd\") pod \"ceilometer-0\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.940337 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deefee78-9e75-44e2-b72a-08aeb8b97ac4-log-httpd\") pod \"ceilometer-0\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.940372 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55fc2\" (UniqueName: \"kubernetes.io/projected/deefee78-9e75-44e2-b72a-08aeb8b97ac4-kube-api-access-55fc2\") pod \"ceilometer-0\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.940434 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-config-data\") pod \"ceilometer-0\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.940586 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-scripts\") pod \"ceilometer-0\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.940619 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.941693 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deefee78-9e75-44e2-b72a-08aeb8b97ac4-run-httpd\") pod \"ceilometer-0\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.941878 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deefee78-9e75-44e2-b72a-08aeb8b97ac4-log-httpd\") pod \"ceilometer-0\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.944623 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.946606 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.946840 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-config-data\") pod \"ceilometer-0\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.947432 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-scripts\") pod \"ceilometer-0\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " pod="openstack/ceilometer-0" Dec 27 08:05:24 crc kubenswrapper[4934]: I1227 08:05:24.970859 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55fc2\" (UniqueName: \"kubernetes.io/projected/deefee78-9e75-44e2-b72a-08aeb8b97ac4-kube-api-access-55fc2\") pod \"ceilometer-0\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " pod="openstack/ceilometer-0" Dec 27 08:05:25 crc kubenswrapper[4934]: I1227 08:05:25.058747 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:05:25 crc kubenswrapper[4934]: I1227 08:05:25.423598 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-76d79bd746-4mrv6" event={"ID":"00fe7217-c021-410c-9fca-36b69e513e92","Type":"ContainerStarted","Data":"b328f82df760cf8a89faa98bcfe54a93e89d6497d5a55e86162969eae8f96216"} Dec 27 08:05:25 crc kubenswrapper[4934]: I1227 08:05:25.432726 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-654859b66d-km642" event={"ID":"700034e0-f221-4bca-a973-a6d49c7dc406","Type":"ContainerStarted","Data":"a65eff45da80910a20b649000a7204ed68ff3e5d08bf8c1dd4a74d55c956682f"} Dec 27 08:05:25 crc kubenswrapper[4934]: I1227 08:05:25.460655 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6bdc8c9885-7z5pz" event={"ID":"2c8feac2-c2ee-484c-baed-052bd7c0b468","Type":"ContainerStarted","Data":"e9bfb9c690528fcbf2b4f8d9a634e42f80fbf609c10f014e8dd8b79e094981af"} Dec 27 08:05:25 crc kubenswrapper[4934]: I1227 08:05:25.460706 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6bdc8c9885-7z5pz" event={"ID":"2c8feac2-c2ee-484c-baed-052bd7c0b468","Type":"ContainerStarted","Data":"3dc9a8c016fe3a231a4ad11a2d12e3bbec94ea7f70c4f3e3924310c286633dd3"} Dec 27 08:05:25 crc kubenswrapper[4934]: I1227 08:05:25.462756 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-6bdc8c9885-7z5pz" Dec 27 08:05:25 crc kubenswrapper[4934]: I1227 08:05:25.490746 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-654859b66d-km642" podStartSLOduration=4.178719068 podStartE2EDuration="8.490719097s" podCreationTimestamp="2025-12-27 08:05:17 +0000 UTC" firstStartedPulling="2025-12-27 08:05:18.297637563 +0000 UTC m=+1379.118078157" lastFinishedPulling="2025-12-27 08:05:22.609637592 +0000 UTC m=+1383.430078186" observedRunningTime="2025-12-27 08:05:25.457212431 +0000 UTC m=+1386.277653025" watchObservedRunningTime="2025-12-27 08:05:25.490719097 +0000 UTC m=+1386.311159691" Dec 27 08:05:25 crc kubenswrapper[4934]: I1227 08:05:25.508622 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48fe2613-0255-4c5c-9b4e-7b26e4360e1e" path="/var/lib/kubelet/pods/48fe2613-0255-4c5c-9b4e-7b26e4360e1e/volumes" Dec 27 08:05:25 crc kubenswrapper[4934]: I1227 08:05:25.512672 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cef13c6c-b21d-45f8-83be-625000146643" path="/var/lib/kubelet/pods/cef13c6c-b21d-45f8-83be-625000146643/volumes" Dec 27 08:05:25 crc kubenswrapper[4934]: I1227 08:05:25.515845 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-859445d775-g44pr" event={"ID":"3f083ef4-9928-400b-969e-1df70b3e9475","Type":"ContainerStarted","Data":"0aec2eb3f0f44bccb91dafa902f07a2aca74598a35649acc5f6a2233d7c70f1e"} Dec 27 08:05:25 crc kubenswrapper[4934]: I1227 08:05:25.519103 4934 generic.go:334] "Generic (PLEG): container finished" podID="7fcd6ab4-3c77-4db4-89b4-a4043523e6b4" containerID="d076109c5b85269e36786feaa1e1e475778b3a2362e59ef25047a167992b795f" exitCode=0 Dec 27 08:05:25 crc kubenswrapper[4934]: I1227 08:05:25.519382 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" event={"ID":"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4","Type":"ContainerDied","Data":"d076109c5b85269e36786feaa1e1e475778b3a2362e59ef25047a167992b795f"} Dec 27 08:05:25 crc kubenswrapper[4934]: I1227 08:05:25.530040 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5cc8cdc7b7-mmj42" event={"ID":"a23b3091-07c0-4617-9fb9-de876bed5e12","Type":"ContainerStarted","Data":"df28844f078156278c57e2df6e32ed3ebf77c1d7b8ad2afa7d23b4993f155d80"} Dec 27 08:05:25 crc kubenswrapper[4934]: I1227 08:05:25.549060 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-6bdc8c9885-7z5pz" podStartSLOduration=3.549033814 podStartE2EDuration="3.549033814s" podCreationTimestamp="2025-12-27 08:05:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:05:25.480688219 +0000 UTC m=+1386.301128823" watchObservedRunningTime="2025-12-27 08:05:25.549033814 +0000 UTC m=+1386.369474408" Dec 27 08:05:25 crc kubenswrapper[4934]: I1227 08:05:25.553560 4934 generic.go:334] "Generic (PLEG): container finished" podID="ee5d7a6d-2acf-4cfa-b028-de1c3a20c546" containerID="c4c45b7d7f63de8bfeca8e55a5b3a1096dd3359a49a71273e20ab35c776b7dc5" exitCode=0 Dec 27 08:05:25 crc kubenswrapper[4934]: I1227 08:05:25.554254 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" event={"ID":"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546","Type":"ContainerDied","Data":"c4c45b7d7f63de8bfeca8e55a5b3a1096dd3359a49a71273e20ab35c776b7dc5"} Dec 27 08:05:25 crc kubenswrapper[4934]: I1227 08:05:25.554331 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" event={"ID":"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546","Type":"ContainerStarted","Data":"ecaa5af55972576661cc763665718fa1c11cce6d34ed7ab546a465a3f6650dbc"} Dec 27 08:05:25 crc kubenswrapper[4934]: I1227 08:05:25.592684 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-859445d775-g44pr" podStartSLOduration=4.031056017 podStartE2EDuration="8.59266672s" podCreationTimestamp="2025-12-27 08:05:17 +0000 UTC" firstStartedPulling="2025-12-27 08:05:18.047056956 +0000 UTC m=+1378.867497550" lastFinishedPulling="2025-12-27 08:05:22.608667659 +0000 UTC m=+1383.429108253" observedRunningTime="2025-12-27 08:05:25.516648046 +0000 UTC m=+1386.337088640" watchObservedRunningTime="2025-12-27 08:05:25.59266672 +0000 UTC m=+1386.413107314" Dec 27 08:05:25 crc kubenswrapper[4934]: I1227 08:05:25.756436 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:05:25 crc kubenswrapper[4934]: W1227 08:05:25.926229 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddeefee78_9e75_44e2_b72a_08aeb8b97ac4.slice/crio-30c22cd6fb03d572171d77615c3621800414a0c1cae613fb9f5dcd05e6d2a7f4 WatchSource:0}: Error finding container 30c22cd6fb03d572171d77615c3621800414a0c1cae613fb9f5dcd05e6d2a7f4: Status 404 returned error can't find the container with id 30c22cd6fb03d572171d77615c3621800414a0c1cae613fb9f5dcd05e6d2a7f4 Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.203850 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.288010 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-config\") pod \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.288189 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-ovsdbserver-nb\") pod \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.288230 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-dns-swift-storage-0\") pod \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.288317 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92tfz\" (UniqueName: \"kubernetes.io/projected/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-kube-api-access-92tfz\") pod \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.288363 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-dns-svc\") pod \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.288412 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-ovsdbserver-sb\") pod \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\" (UID: \"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4\") " Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.295238 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-kube-api-access-92tfz" (OuterVolumeSpecName: "kube-api-access-92tfz") pod "7fcd6ab4-3c77-4db4-89b4-a4043523e6b4" (UID: "7fcd6ab4-3c77-4db4-89b4-a4043523e6b4"). InnerVolumeSpecName "kube-api-access-92tfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.354961 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7fcd6ab4-3c77-4db4-89b4-a4043523e6b4" (UID: "7fcd6ab4-3c77-4db4-89b4-a4043523e6b4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.394657 4934 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.394686 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92tfz\" (UniqueName: \"kubernetes.io/projected/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-kube-api-access-92tfz\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.417591 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7fcd6ab4-3c77-4db4-89b4-a4043523e6b4" (UID: "7fcd6ab4-3c77-4db4-89b4-a4043523e6b4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.419134 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-config" (OuterVolumeSpecName: "config") pod "7fcd6ab4-3c77-4db4-89b4-a4043523e6b4" (UID: "7fcd6ab4-3c77-4db4-89b4-a4043523e6b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.510833 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.510868 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.532762 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7fcd6ab4-3c77-4db4-89b4-a4043523e6b4" (UID: "7fcd6ab4-3c77-4db4-89b4-a4043523e6b4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.552961 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7fcd6ab4-3c77-4db4-89b4-a4043523e6b4" (UID: "7fcd6ab4-3c77-4db4-89b4-a4043523e6b4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.592360 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-76886c86d8-c6mdr" Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.605028 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" event={"ID":"7fcd6ab4-3c77-4db4-89b4-a4043523e6b4","Type":"ContainerDied","Data":"0ccb34a6fc7e56df39a724b71be390590ebe6c8f5b2f399183e44e9942596ba1"} Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.605159 4934 scope.go:117] "RemoveContainer" containerID="d076109c5b85269e36786feaa1e1e475778b3a2362e59ef25047a167992b795f" Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.605336 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-5gt6v" Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.612644 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.612668 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.633319 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-76d79bd746-4mrv6" event={"ID":"00fe7217-c021-410c-9fca-36b69e513e92","Type":"ContainerStarted","Data":"e7b6923fa5e5601975d5b522c8e28aae4cd9aadb89a11e6af6680dca45e217b6"} Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.634643 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.634667 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.638701 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deefee78-9e75-44e2-b72a-08aeb8b97ac4","Type":"ContainerStarted","Data":"30c22cd6fb03d572171d77615c3621800414a0c1cae613fb9f5dcd05e6d2a7f4"} Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.689037 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-76d79bd746-4mrv6" podStartSLOduration=6.689020717 podStartE2EDuration="6.689020717s" podCreationTimestamp="2025-12-27 08:05:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:05:26.680351054 +0000 UTC m=+1387.500791648" watchObservedRunningTime="2025-12-27 08:05:26.689020717 +0000 UTC m=+1387.509461311" Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.702334 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"34c0fb07-e3e3-45f0-b785-5c981a771ceb","Type":"ContainerStarted","Data":"6ba9e0de8f593334de9c3a3f840f5b826c635cfefee0191b276a6548829c4f4b"} Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.779851 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.813582 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-5gt6v"] Dec 27 08:05:26 crc kubenswrapper[4934]: I1227 08:05:26.836967 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-5gt6v"] Dec 27 08:05:27 crc kubenswrapper[4934]: I1227 08:05:27.479466 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fcd6ab4-3c77-4db4-89b4-a4043523e6b4" path="/var/lib/kubelet/pods/7fcd6ab4-3c77-4db4-89b4-a4043523e6b4/volumes" Dec 27 08:05:28 crc kubenswrapper[4934]: I1227 08:05:28.304113 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-c689587cc-6glpp" Dec 27 08:05:28 crc kubenswrapper[4934]: I1227 08:05:28.368682 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-76886c86d8-c6mdr"] Dec 27 08:05:28 crc kubenswrapper[4934]: I1227 08:05:28.368909 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-76886c86d8-c6mdr" podUID="c67ffca3-2539-4344-848e-e251e564f42b" containerName="neutron-api" containerID="cri-o://267c584dc50d95618fa637ec47ff3c93f9e2ea0d158207d0591b2838c4afa90d" gracePeriod=30 Dec 27 08:05:28 crc kubenswrapper[4934]: I1227 08:05:28.369013 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-76886c86d8-c6mdr" podUID="c67ffca3-2539-4344-848e-e251e564f42b" containerName="neutron-httpd" containerID="cri-o://67e0ab891a2e8f56faeae88e28a6fca9233b0b966fcab3583810609696d05004" gracePeriod=30 Dec 27 08:05:28 crc kubenswrapper[4934]: I1227 08:05:28.613409 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-58d7546b8b-lqdtc" podUID="327eb20e-e905-48a6-99b8-9209813e3330" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.202:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 08:05:28 crc kubenswrapper[4934]: I1227 08:05:28.613537 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-58d7546b8b-lqdtc" podUID="327eb20e-e905-48a6-99b8-9209813e3330" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.202:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 08:05:28 crc kubenswrapper[4934]: I1227 08:05:28.730337 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"34c0fb07-e3e3-45f0-b785-5c981a771ceb","Type":"ContainerStarted","Data":"958bfd98687130b5348a39978c055d6505574984ccbf8cb3b6e95825cf5dc1c7"} Dec 27 08:05:28 crc kubenswrapper[4934]: I1227 08:05:28.730430 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="34c0fb07-e3e3-45f0-b785-5c981a771ceb" containerName="cinder-api-log" containerID="cri-o://6ba9e0de8f593334de9c3a3f840f5b826c635cfefee0191b276a6548829c4f4b" gracePeriod=30 Dec 27 08:05:28 crc kubenswrapper[4934]: I1227 08:05:28.730574 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="34c0fb07-e3e3-45f0-b785-5c981a771ceb" containerName="cinder-api" containerID="cri-o://958bfd98687130b5348a39978c055d6505574984ccbf8cb3b6e95825cf5dc1c7" gracePeriod=30 Dec 27 08:05:28 crc kubenswrapper[4934]: I1227 08:05:28.730775 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 27 08:05:28 crc kubenswrapper[4934]: I1227 08:05:28.739108 4934 generic.go:334] "Generic (PLEG): container finished" podID="c67ffca3-2539-4344-848e-e251e564f42b" containerID="67e0ab891a2e8f56faeae88e28a6fca9233b0b966fcab3583810609696d05004" exitCode=0 Dec 27 08:05:28 crc kubenswrapper[4934]: I1227 08:05:28.739373 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76886c86d8-c6mdr" event={"ID":"c67ffca3-2539-4344-848e-e251e564f42b","Type":"ContainerDied","Data":"67e0ab891a2e8f56faeae88e28a6fca9233b0b966fcab3583810609696d05004"} Dec 27 08:05:28 crc kubenswrapper[4934]: I1227 08:05:28.741717 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" event={"ID":"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546","Type":"ContainerStarted","Data":"91212881dafecfe654aa3a58c9f80410ec17d93fb18b80c38eb052ad42b89c8a"} Dec 27 08:05:28 crc kubenswrapper[4934]: I1227 08:05:28.741895 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:28 crc kubenswrapper[4934]: I1227 08:05:28.757202 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=7.757184432 podStartE2EDuration="7.757184432s" podCreationTimestamp="2025-12-27 08:05:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:05:28.753612684 +0000 UTC m=+1389.574053298" watchObservedRunningTime="2025-12-27 08:05:28.757184432 +0000 UTC m=+1389.577625026" Dec 27 08:05:28 crc kubenswrapper[4934]: I1227 08:05:28.791552 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" podStartSLOduration=6.791529579 podStartE2EDuration="6.791529579s" podCreationTimestamp="2025-12-27 08:05:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:05:28.777590466 +0000 UTC m=+1389.598031080" watchObservedRunningTime="2025-12-27 08:05:28.791529579 +0000 UTC m=+1389.611970173" Dec 27 08:05:29 crc kubenswrapper[4934]: I1227 08:05:29.730839 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:29 crc kubenswrapper[4934]: I1227 08:05:29.746938 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-788c8bbcbc-rtffl" Dec 27 08:05:29 crc kubenswrapper[4934]: I1227 08:05:29.755435 4934 generic.go:334] "Generic (PLEG): container finished" podID="34c0fb07-e3e3-45f0-b785-5c981a771ceb" containerID="6ba9e0de8f593334de9c3a3f840f5b826c635cfefee0191b276a6548829c4f4b" exitCode=143 Dec 27 08:05:29 crc kubenswrapper[4934]: I1227 08:05:29.755986 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"34c0fb07-e3e3-45f0-b785-5c981a771ceb","Type":"ContainerDied","Data":"6ba9e0de8f593334de9c3a3f840f5b826c635cfefee0191b276a6548829c4f4b"} Dec 27 08:05:30 crc kubenswrapper[4934]: I1227 08:05:30.772706 4934 generic.go:334] "Generic (PLEG): container finished" podID="c67ffca3-2539-4344-848e-e251e564f42b" containerID="267c584dc50d95618fa637ec47ff3c93f9e2ea0d158207d0591b2838c4afa90d" exitCode=0 Dec 27 08:05:30 crc kubenswrapper[4934]: I1227 08:05:30.773045 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76886c86d8-c6mdr" event={"ID":"c67ffca3-2539-4344-848e-e251e564f42b","Type":"ContainerDied","Data":"267c584dc50d95618fa637ec47ff3c93f9e2ea0d158207d0591b2838c4afa90d"} Dec 27 08:05:30 crc kubenswrapper[4934]: I1227 08:05:30.886250 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.088878 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.358180 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-6f98b67c6-55vgd"] Dec 27 08:05:32 crc kubenswrapper[4934]: E1227 08:05:32.358739 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fcd6ab4-3c77-4db4-89b4-a4043523e6b4" containerName="init" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.358754 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fcd6ab4-3c77-4db4-89b4-a4043523e6b4" containerName="init" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.358942 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fcd6ab4-3c77-4db4-89b4-a4043523e6b4" containerName="init" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.359763 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6f98b67c6-55vgd" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.400305 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6f98b67c6-55vgd"] Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.424261 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-7f59f88684-gxs5t"] Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.426932 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7f59f88684-gxs5t" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.470970 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7f59f88684-gxs5t"] Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.482153 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-db89fbccf-w9xvz"] Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.483844 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-db89fbccf-w9xvz" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.488106 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1515c241-e28d-49d7-a3f3-790aa082251c-config-data-custom\") pod \"heat-engine-6f98b67c6-55vgd\" (UID: \"1515c241-e28d-49d7-a3f3-790aa082251c\") " pod="openstack/heat-engine-6f98b67c6-55vgd" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.488197 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8gtw\" (UniqueName: \"kubernetes.io/projected/1515c241-e28d-49d7-a3f3-790aa082251c-kube-api-access-j8gtw\") pod \"heat-engine-6f98b67c6-55vgd\" (UID: \"1515c241-e28d-49d7-a3f3-790aa082251c\") " pod="openstack/heat-engine-6f98b67c6-55vgd" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.488446 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1515c241-e28d-49d7-a3f3-790aa082251c-combined-ca-bundle\") pod \"heat-engine-6f98b67c6-55vgd\" (UID: \"1515c241-e28d-49d7-a3f3-790aa082251c\") " pod="openstack/heat-engine-6f98b67c6-55vgd" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.488493 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1515c241-e28d-49d7-a3f3-790aa082251c-config-data\") pod \"heat-engine-6f98b67c6-55vgd\" (UID: \"1515c241-e28d-49d7-a3f3-790aa082251c\") " pod="openstack/heat-engine-6f98b67c6-55vgd" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.507382 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-db89fbccf-w9xvz"] Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.590829 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7237f7ae-a705-41af-ace6-27bafbba974d-config-data-custom\") pod \"heat-api-db89fbccf-w9xvz\" (UID: \"7237f7ae-a705-41af-ace6-27bafbba974d\") " pod="openstack/heat-api-db89fbccf-w9xvz" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.590939 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-config-data-custom\") pod \"heat-cfnapi-7f59f88684-gxs5t\" (UID: \"1182d8a5-1924-45e8-b13f-18fc3b25cf9c\") " pod="openstack/heat-cfnapi-7f59f88684-gxs5t" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.591053 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1515c241-e28d-49d7-a3f3-790aa082251c-config-data-custom\") pod \"heat-engine-6f98b67c6-55vgd\" (UID: \"1515c241-e28d-49d7-a3f3-790aa082251c\") " pod="openstack/heat-engine-6f98b67c6-55vgd" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.591143 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8gtw\" (UniqueName: \"kubernetes.io/projected/1515c241-e28d-49d7-a3f3-790aa082251c-kube-api-access-j8gtw\") pod \"heat-engine-6f98b67c6-55vgd\" (UID: \"1515c241-e28d-49d7-a3f3-790aa082251c\") " pod="openstack/heat-engine-6f98b67c6-55vgd" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.591272 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7237f7ae-a705-41af-ace6-27bafbba974d-combined-ca-bundle\") pod \"heat-api-db89fbccf-w9xvz\" (UID: \"7237f7ae-a705-41af-ace6-27bafbba974d\") " pod="openstack/heat-api-db89fbccf-w9xvz" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.591358 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7237f7ae-a705-41af-ace6-27bafbba974d-config-data\") pod \"heat-api-db89fbccf-w9xvz\" (UID: \"7237f7ae-a705-41af-ace6-27bafbba974d\") " pod="openstack/heat-api-db89fbccf-w9xvz" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.591393 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-combined-ca-bundle\") pod \"heat-cfnapi-7f59f88684-gxs5t\" (UID: \"1182d8a5-1924-45e8-b13f-18fc3b25cf9c\") " pod="openstack/heat-cfnapi-7f59f88684-gxs5t" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.591473 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-config-data\") pod \"heat-cfnapi-7f59f88684-gxs5t\" (UID: \"1182d8a5-1924-45e8-b13f-18fc3b25cf9c\") " pod="openstack/heat-cfnapi-7f59f88684-gxs5t" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.591507 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jn7g\" (UniqueName: \"kubernetes.io/projected/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-kube-api-access-5jn7g\") pod \"heat-cfnapi-7f59f88684-gxs5t\" (UID: \"1182d8a5-1924-45e8-b13f-18fc3b25cf9c\") " pod="openstack/heat-cfnapi-7f59f88684-gxs5t" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.591594 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sxd6\" (UniqueName: \"kubernetes.io/projected/7237f7ae-a705-41af-ace6-27bafbba974d-kube-api-access-8sxd6\") pod \"heat-api-db89fbccf-w9xvz\" (UID: \"7237f7ae-a705-41af-ace6-27bafbba974d\") " pod="openstack/heat-api-db89fbccf-w9xvz" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.591642 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1515c241-e28d-49d7-a3f3-790aa082251c-combined-ca-bundle\") pod \"heat-engine-6f98b67c6-55vgd\" (UID: \"1515c241-e28d-49d7-a3f3-790aa082251c\") " pod="openstack/heat-engine-6f98b67c6-55vgd" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.591679 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1515c241-e28d-49d7-a3f3-790aa082251c-config-data\") pod \"heat-engine-6f98b67c6-55vgd\" (UID: \"1515c241-e28d-49d7-a3f3-790aa082251c\") " pod="openstack/heat-engine-6f98b67c6-55vgd" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.608649 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1515c241-e28d-49d7-a3f3-790aa082251c-config-data-custom\") pod \"heat-engine-6f98b67c6-55vgd\" (UID: \"1515c241-e28d-49d7-a3f3-790aa082251c\") " pod="openstack/heat-engine-6f98b67c6-55vgd" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.611833 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1515c241-e28d-49d7-a3f3-790aa082251c-combined-ca-bundle\") pod \"heat-engine-6f98b67c6-55vgd\" (UID: \"1515c241-e28d-49d7-a3f3-790aa082251c\") " pod="openstack/heat-engine-6f98b67c6-55vgd" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.616891 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8gtw\" (UniqueName: \"kubernetes.io/projected/1515c241-e28d-49d7-a3f3-790aa082251c-kube-api-access-j8gtw\") pod \"heat-engine-6f98b67c6-55vgd\" (UID: \"1515c241-e28d-49d7-a3f3-790aa082251c\") " pod="openstack/heat-engine-6f98b67c6-55vgd" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.643617 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1515c241-e28d-49d7-a3f3-790aa082251c-config-data\") pod \"heat-engine-6f98b67c6-55vgd\" (UID: \"1515c241-e28d-49d7-a3f3-790aa082251c\") " pod="openstack/heat-engine-6f98b67c6-55vgd" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.703507 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sxd6\" (UniqueName: \"kubernetes.io/projected/7237f7ae-a705-41af-ace6-27bafbba974d-kube-api-access-8sxd6\") pod \"heat-api-db89fbccf-w9xvz\" (UID: \"7237f7ae-a705-41af-ace6-27bafbba974d\") " pod="openstack/heat-api-db89fbccf-w9xvz" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.703680 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7237f7ae-a705-41af-ace6-27bafbba974d-config-data-custom\") pod \"heat-api-db89fbccf-w9xvz\" (UID: \"7237f7ae-a705-41af-ace6-27bafbba974d\") " pod="openstack/heat-api-db89fbccf-w9xvz" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.703729 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-config-data-custom\") pod \"heat-cfnapi-7f59f88684-gxs5t\" (UID: \"1182d8a5-1924-45e8-b13f-18fc3b25cf9c\") " pod="openstack/heat-cfnapi-7f59f88684-gxs5t" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.703884 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7237f7ae-a705-41af-ace6-27bafbba974d-combined-ca-bundle\") pod \"heat-api-db89fbccf-w9xvz\" (UID: \"7237f7ae-a705-41af-ace6-27bafbba974d\") " pod="openstack/heat-api-db89fbccf-w9xvz" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.703933 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7237f7ae-a705-41af-ace6-27bafbba974d-config-data\") pod \"heat-api-db89fbccf-w9xvz\" (UID: \"7237f7ae-a705-41af-ace6-27bafbba974d\") " pod="openstack/heat-api-db89fbccf-w9xvz" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.703956 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-combined-ca-bundle\") pod \"heat-cfnapi-7f59f88684-gxs5t\" (UID: \"1182d8a5-1924-45e8-b13f-18fc3b25cf9c\") " pod="openstack/heat-cfnapi-7f59f88684-gxs5t" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.704008 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-config-data\") pod \"heat-cfnapi-7f59f88684-gxs5t\" (UID: \"1182d8a5-1924-45e8-b13f-18fc3b25cf9c\") " pod="openstack/heat-cfnapi-7f59f88684-gxs5t" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.704037 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jn7g\" (UniqueName: \"kubernetes.io/projected/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-kube-api-access-5jn7g\") pod \"heat-cfnapi-7f59f88684-gxs5t\" (UID: \"1182d8a5-1924-45e8-b13f-18fc3b25cf9c\") " pod="openstack/heat-cfnapi-7f59f88684-gxs5t" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.704582 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6f98b67c6-55vgd" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.714997 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-config-data-custom\") pod \"heat-cfnapi-7f59f88684-gxs5t\" (UID: \"1182d8a5-1924-45e8-b13f-18fc3b25cf9c\") " pod="openstack/heat-cfnapi-7f59f88684-gxs5t" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.715222 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7237f7ae-a705-41af-ace6-27bafbba974d-config-data\") pod \"heat-api-db89fbccf-w9xvz\" (UID: \"7237f7ae-a705-41af-ace6-27bafbba974d\") " pod="openstack/heat-api-db89fbccf-w9xvz" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.717361 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-combined-ca-bundle\") pod \"heat-cfnapi-7f59f88684-gxs5t\" (UID: \"1182d8a5-1924-45e8-b13f-18fc3b25cf9c\") " pod="openstack/heat-cfnapi-7f59f88684-gxs5t" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.719009 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7237f7ae-a705-41af-ace6-27bafbba974d-config-data-custom\") pod \"heat-api-db89fbccf-w9xvz\" (UID: \"7237f7ae-a705-41af-ace6-27bafbba974d\") " pod="openstack/heat-api-db89fbccf-w9xvz" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.719627 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7237f7ae-a705-41af-ace6-27bafbba974d-combined-ca-bundle\") pod \"heat-api-db89fbccf-w9xvz\" (UID: \"7237f7ae-a705-41af-ace6-27bafbba974d\") " pod="openstack/heat-api-db89fbccf-w9xvz" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.719972 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-config-data\") pod \"heat-cfnapi-7f59f88684-gxs5t\" (UID: \"1182d8a5-1924-45e8-b13f-18fc3b25cf9c\") " pod="openstack/heat-cfnapi-7f59f88684-gxs5t" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.723439 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jn7g\" (UniqueName: \"kubernetes.io/projected/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-kube-api-access-5jn7g\") pod \"heat-cfnapi-7f59f88684-gxs5t\" (UID: \"1182d8a5-1924-45e8-b13f-18fc3b25cf9c\") " pod="openstack/heat-cfnapi-7f59f88684-gxs5t" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.738737 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sxd6\" (UniqueName: \"kubernetes.io/projected/7237f7ae-a705-41af-ace6-27bafbba974d-kube-api-access-8sxd6\") pod \"heat-api-db89fbccf-w9xvz\" (UID: \"7237f7ae-a705-41af-ace6-27bafbba974d\") " pod="openstack/heat-api-db89fbccf-w9xvz" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.790639 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7f59f88684-gxs5t" Dec 27 08:05:32 crc kubenswrapper[4934]: I1227 08:05:32.834664 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-db89fbccf-w9xvz" Dec 27 08:05:34 crc kubenswrapper[4934]: I1227 08:05:34.484240 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:34 crc kubenswrapper[4934]: I1227 08:05:34.504606 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-76d79bd746-4mrv6" Dec 27 08:05:34 crc kubenswrapper[4934]: I1227 08:05:34.583485 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-58d7546b8b-lqdtc"] Dec 27 08:05:34 crc kubenswrapper[4934]: I1227 08:05:34.583691 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-58d7546b8b-lqdtc" podUID="327eb20e-e905-48a6-99b8-9209813e3330" containerName="barbican-api-log" containerID="cri-o://f7c0182c802c5478d3e949a8b8cfad73ed9ed473d0b1b437a975df41dd930ac0" gracePeriod=30 Dec 27 08:05:34 crc kubenswrapper[4934]: I1227 08:05:34.583875 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-58d7546b8b-lqdtc" podUID="327eb20e-e905-48a6-99b8-9209813e3330" containerName="barbican-api" containerID="cri-o://440c2ffcdd51fdd7c00a40b0c3598aa5f6961ca95292c78d0f54a9362030242c" gracePeriod=30 Dec 27 08:05:34 crc kubenswrapper[4934]: I1227 08:05:34.604302 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-58d7546b8b-lqdtc" podUID="327eb20e-e905-48a6-99b8-9209813e3330" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.202:9311/healthcheck\": EOF" Dec 27 08:05:34 crc kubenswrapper[4934]: I1227 08:05:34.830045 4934 generic.go:334] "Generic (PLEG): container finished" podID="327eb20e-e905-48a6-99b8-9209813e3330" containerID="f7c0182c802c5478d3e949a8b8cfad73ed9ed473d0b1b437a975df41dd930ac0" exitCode=143 Dec 27 08:05:34 crc kubenswrapper[4934]: I1227 08:05:34.830468 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58d7546b8b-lqdtc" event={"ID":"327eb20e-e905-48a6-99b8-9209813e3330","Type":"ContainerDied","Data":"f7c0182c802c5478d3e949a8b8cfad73ed9ed473d0b1b437a975df41dd930ac0"} Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.592552 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5cc8cdc7b7-mmj42"] Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.604218 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-f7fc9c49c-vpz65"] Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.615737 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-785dfb69f-wftff"] Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.617535 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.621068 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.621252 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.632218 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-cf699bf96-qsv9c"] Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.633738 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.638882 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.639213 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.645797 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-cf699bf96-qsv9c"] Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.661652 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-785dfb69f-wftff"] Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.676926 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-public-tls-certs\") pod \"heat-api-785dfb69f-wftff\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.677065 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-combined-ca-bundle\") pod \"heat-api-785dfb69f-wftff\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.677158 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-config-data-custom\") pod \"heat-api-785dfb69f-wftff\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.677225 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s24vm\" (UniqueName: \"kubernetes.io/projected/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-kube-api-access-s24vm\") pod \"heat-api-785dfb69f-wftff\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.677274 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-internal-tls-certs\") pod \"heat-api-785dfb69f-wftff\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.677350 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-config-data\") pod \"heat-api-785dfb69f-wftff\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.779458 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-internal-tls-certs\") pod \"heat-api-785dfb69f-wftff\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.779780 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rc8z\" (UniqueName: \"kubernetes.io/projected/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-kube-api-access-9rc8z\") pod \"heat-cfnapi-cf699bf96-qsv9c\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.779813 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-config-data-custom\") pod \"heat-cfnapi-cf699bf96-qsv9c\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.779833 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-combined-ca-bundle\") pod \"heat-cfnapi-cf699bf96-qsv9c\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.779891 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-config-data\") pod \"heat-api-785dfb69f-wftff\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.779923 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-public-tls-certs\") pod \"heat-api-785dfb69f-wftff\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.779962 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-internal-tls-certs\") pod \"heat-cfnapi-cf699bf96-qsv9c\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.780055 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-combined-ca-bundle\") pod \"heat-api-785dfb69f-wftff\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.780121 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-config-data\") pod \"heat-cfnapi-cf699bf96-qsv9c\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.780192 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-config-data-custom\") pod \"heat-api-785dfb69f-wftff\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.780263 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s24vm\" (UniqueName: \"kubernetes.io/projected/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-kube-api-access-s24vm\") pod \"heat-api-785dfb69f-wftff\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.780303 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-public-tls-certs\") pod \"heat-cfnapi-cf699bf96-qsv9c\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.785692 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-public-tls-certs\") pod \"heat-api-785dfb69f-wftff\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.786220 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-config-data\") pod \"heat-api-785dfb69f-wftff\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.786961 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-combined-ca-bundle\") pod \"heat-api-785dfb69f-wftff\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.787598 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-internal-tls-certs\") pod \"heat-api-785dfb69f-wftff\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.788007 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-config-data-custom\") pod \"heat-api-785dfb69f-wftff\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.801141 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s24vm\" (UniqueName: \"kubernetes.io/projected/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-kube-api-access-s24vm\") pod \"heat-api-785dfb69f-wftff\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.882158 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-public-tls-certs\") pod \"heat-cfnapi-cf699bf96-qsv9c\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.882460 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rc8z\" (UniqueName: \"kubernetes.io/projected/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-kube-api-access-9rc8z\") pod \"heat-cfnapi-cf699bf96-qsv9c\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.882565 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-config-data-custom\") pod \"heat-cfnapi-cf699bf96-qsv9c\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.882679 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-combined-ca-bundle\") pod \"heat-cfnapi-cf699bf96-qsv9c\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.882857 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-internal-tls-certs\") pod \"heat-cfnapi-cf699bf96-qsv9c\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.882984 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-config-data\") pod \"heat-cfnapi-cf699bf96-qsv9c\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.887835 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-config-data-custom\") pod \"heat-cfnapi-cf699bf96-qsv9c\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.888253 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-combined-ca-bundle\") pod \"heat-cfnapi-cf699bf96-qsv9c\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.889707 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-internal-tls-certs\") pod \"heat-cfnapi-cf699bf96-qsv9c\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.889856 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-config-data\") pod \"heat-cfnapi-cf699bf96-qsv9c\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.890567 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-public-tls-certs\") pod \"heat-cfnapi-cf699bf96-qsv9c\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.901889 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rc8z\" (UniqueName: \"kubernetes.io/projected/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-kube-api-access-9rc8z\") pod \"heat-cfnapi-cf699bf96-qsv9c\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.957630 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:35 crc kubenswrapper[4934]: I1227 08:05:35.967945 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:37 crc kubenswrapper[4934]: I1227 08:05:37.869274 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:05:37 crc kubenswrapper[4934]: I1227 08:05:37.927934 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-76g9s"] Dec 27 08:05:37 crc kubenswrapper[4934]: I1227 08:05:37.929167 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-76g9s" podUID="ea0027c6-81c4-4d65-96a6-43eb035d2043" containerName="dnsmasq-dns" containerID="cri-o://416ae62c615270f259b28149ef9568a4f45c93acc55599fbdfd7470a080632e2" gracePeriod=10 Dec 27 08:05:38 crc kubenswrapper[4934]: I1227 08:05:38.062757 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-58d7546b8b-lqdtc" podUID="327eb20e-e905-48a6-99b8-9209813e3330" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.202:9311/healthcheck\": read tcp 10.217.0.2:35522->10.217.0.202:9311: read: connection reset by peer" Dec 27 08:05:38 crc kubenswrapper[4934]: I1227 08:05:38.062784 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-58d7546b8b-lqdtc" podUID="327eb20e-e905-48a6-99b8-9209813e3330" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.202:9311/healthcheck\": read tcp 10.217.0.2:35532->10.217.0.202:9311: read: connection reset by peer" Dec 27 08:05:38 crc kubenswrapper[4934]: I1227 08:05:38.884512 4934 generic.go:334] "Generic (PLEG): container finished" podID="327eb20e-e905-48a6-99b8-9209813e3330" containerID="440c2ffcdd51fdd7c00a40b0c3598aa5f6961ca95292c78d0f54a9362030242c" exitCode=0 Dec 27 08:05:38 crc kubenswrapper[4934]: I1227 08:05:38.884811 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58d7546b8b-lqdtc" event={"ID":"327eb20e-e905-48a6-99b8-9209813e3330","Type":"ContainerDied","Data":"440c2ffcdd51fdd7c00a40b0c3598aa5f6961ca95292c78d0f54a9362030242c"} Dec 27 08:05:38 crc kubenswrapper[4934]: I1227 08:05:38.888453 4934 generic.go:334] "Generic (PLEG): container finished" podID="ea0027c6-81c4-4d65-96a6-43eb035d2043" containerID="416ae62c615270f259b28149ef9568a4f45c93acc55599fbdfd7470a080632e2" exitCode=0 Dec 27 08:05:38 crc kubenswrapper[4934]: I1227 08:05:38.888477 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-76g9s" event={"ID":"ea0027c6-81c4-4d65-96a6-43eb035d2043","Type":"ContainerDied","Data":"416ae62c615270f259b28149ef9568a4f45c93acc55599fbdfd7470a080632e2"} Dec 27 08:05:39 crc kubenswrapper[4934]: I1227 08:05:39.653407 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 27 08:05:39 crc kubenswrapper[4934]: E1227 08:05:39.782318 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified" Dec 27 08:05:39 crc kubenswrapper[4934]: E1227 08:05:39.782484 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstackclient,Image:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,Command:[/bin/sleep],Args:[infinity],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5fdh57hd4h56fhd6h687hfdhd8h574h9chc4h565h68fh555h59ch588h54h58ch544h64dh587h658h685h644hddh64dh575h56h68bh58bh5b4h77q,ValueFrom:nil,},EnvVar{Name:OS_CLOUD,Value:default,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_CA_CERT,Value:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_HOST,Value:metric-storage-prometheus.openstack.svc,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_PORT,Value:9090,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openstack-config,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/cloudrc,SubPath:cloudrc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d7qpf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42401,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42401,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstackclient_openstack(f5eabd4f-a103-4ec0-9c14-7deb0e36fe61): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:05:39 crc kubenswrapper[4934]: E1227 08:05:39.783851 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstackclient" podUID="f5eabd4f-a103-4ec0-9c14-7deb0e36fe61" Dec 27 08:05:39 crc kubenswrapper[4934]: E1227 08:05:39.903244 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified\\\"\"" pod="openstack/openstackclient" podUID="f5eabd4f-a103-4ec0-9c14-7deb0e36fe61" Dec 27 08:05:40 crc kubenswrapper[4934]: E1227 08:05:40.372505 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified" Dec 27 08:05:40 crc kubenswrapper[4934]: E1227 08:05:40.373051 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-cfnapi,Image:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_httpd_setup && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5c8hbdh59ch76h686h685h58h656h5dch68dh65hfch5cfh5d4h5cchddhf5h5c6h5b7h5d7h68dhd9h5b4h558h666h9bh598h5dch5cdh596hd4h688q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:heat-cfnapi-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-custom,ReadOnly:true,MountPath:/etc/heat/heat.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wjc2p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthcheck,Port:{0 8000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:10,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthcheck,Port:{0 8000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:10,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-cfnapi-f7fc9c49c-vpz65_openstack(7b983753-0d99-48aa-82e9-c5a0643cf4c3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:05:40 crc kubenswrapper[4934]: E1227 08:05:40.374361 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-cfnapi-f7fc9c49c-vpz65" podUID="7b983753-0d99-48aa-82e9-c5a0643cf4c3" Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.445001 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.445307 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f994d21a-55df-4ea2-a86d-41d18d10d4d1" containerName="glance-log" containerID="cri-o://5738140d2bba76b7bc6a908b15245ca158c60b158eba7d68b51d42af6c3a07b2" gracePeriod=30 Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.445825 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f994d21a-55df-4ea2-a86d-41d18d10d4d1" containerName="glance-httpd" containerID="cri-o://d16fe7c3850179dc183ac942ae64cb6065521e25fa98f3bc7089f111a9b8c0a1" gracePeriod=30 Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.519852 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.636132 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/327eb20e-e905-48a6-99b8-9209813e3330-config-data-custom\") pod \"327eb20e-e905-48a6-99b8-9209813e3330\" (UID: \"327eb20e-e905-48a6-99b8-9209813e3330\") " Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.636222 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/327eb20e-e905-48a6-99b8-9209813e3330-combined-ca-bundle\") pod \"327eb20e-e905-48a6-99b8-9209813e3330\" (UID: \"327eb20e-e905-48a6-99b8-9209813e3330\") " Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.636366 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/327eb20e-e905-48a6-99b8-9209813e3330-logs\") pod \"327eb20e-e905-48a6-99b8-9209813e3330\" (UID: \"327eb20e-e905-48a6-99b8-9209813e3330\") " Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.636424 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/327eb20e-e905-48a6-99b8-9209813e3330-config-data\") pod \"327eb20e-e905-48a6-99b8-9209813e3330\" (UID: \"327eb20e-e905-48a6-99b8-9209813e3330\") " Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.636513 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fw5vp\" (UniqueName: \"kubernetes.io/projected/327eb20e-e905-48a6-99b8-9209813e3330-kube-api-access-fw5vp\") pod \"327eb20e-e905-48a6-99b8-9209813e3330\" (UID: \"327eb20e-e905-48a6-99b8-9209813e3330\") " Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.637264 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/327eb20e-e905-48a6-99b8-9209813e3330-logs" (OuterVolumeSpecName: "logs") pod "327eb20e-e905-48a6-99b8-9209813e3330" (UID: "327eb20e-e905-48a6-99b8-9209813e3330"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.644440 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/327eb20e-e905-48a6-99b8-9209813e3330-kube-api-access-fw5vp" (OuterVolumeSpecName: "kube-api-access-fw5vp") pod "327eb20e-e905-48a6-99b8-9209813e3330" (UID: "327eb20e-e905-48a6-99b8-9209813e3330"). InnerVolumeSpecName "kube-api-access-fw5vp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.644447 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/327eb20e-e905-48a6-99b8-9209813e3330-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "327eb20e-e905-48a6-99b8-9209813e3330" (UID: "327eb20e-e905-48a6-99b8-9209813e3330"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.667778 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/327eb20e-e905-48a6-99b8-9209813e3330-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "327eb20e-e905-48a6-99b8-9209813e3330" (UID: "327eb20e-e905-48a6-99b8-9209813e3330"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.698634 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/327eb20e-e905-48a6-99b8-9209813e3330-config-data" (OuterVolumeSpecName: "config-data") pod "327eb20e-e905-48a6-99b8-9209813e3330" (UID: "327eb20e-e905-48a6-99b8-9209813e3330"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.739569 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fw5vp\" (UniqueName: \"kubernetes.io/projected/327eb20e-e905-48a6-99b8-9209813e3330-kube-api-access-fw5vp\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.739602 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/327eb20e-e905-48a6-99b8-9209813e3330-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.739612 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/327eb20e-e905-48a6-99b8-9209813e3330-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.739623 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/327eb20e-e905-48a6-99b8-9209813e3330-logs\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.739634 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/327eb20e-e905-48a6-99b8-9209813e3330-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.911869 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58d7546b8b-lqdtc" Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.911856 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58d7546b8b-lqdtc" event={"ID":"327eb20e-e905-48a6-99b8-9209813e3330","Type":"ContainerDied","Data":"6cdbf81a31377bc5e87d4e32aeaa967e4c730346fc3ade778edeb7ee84f4ce3b"} Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.911998 4934 scope.go:117] "RemoveContainer" containerID="440c2ffcdd51fdd7c00a40b0c3598aa5f6961ca95292c78d0f54a9362030242c" Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.920553 4934 generic.go:334] "Generic (PLEG): container finished" podID="f994d21a-55df-4ea2-a86d-41d18d10d4d1" containerID="5738140d2bba76b7bc6a908b15245ca158c60b158eba7d68b51d42af6c3a07b2" exitCode=143 Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.920693 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f994d21a-55df-4ea2-a86d-41d18d10d4d1","Type":"ContainerDied","Data":"5738140d2bba76b7bc6a908b15245ca158c60b158eba7d68b51d42af6c3a07b2"} Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.966588 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-58d7546b8b-lqdtc"] Dec 27 08:05:40 crc kubenswrapper[4934]: I1227 08:05:40.981318 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-58d7546b8b-lqdtc"] Dec 27 08:05:41 crc kubenswrapper[4934]: E1227 08:05:41.035243 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-api:current-podified" Dec 27 08:05:41 crc kubenswrapper[4934]: E1227 08:05:41.035432 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-api,Image:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_httpd_setup && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5ffh5b9h666hf4h5d6h587hdbh98h85h646hd8hfbh5d4h59dh558h5bh66bhd9h5f8h664h86h8bh674h7bh648h64hd4h665h79h684h684h5c8q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:heat-api-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-custom,ReadOnly:true,MountPath:/etc/heat/heat.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pdxpj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthcheck,Port:{0 8004 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:10,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthcheck,Port:{0 8004 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:10,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-api-5cc8cdc7b7-mmj42_openstack(a23b3091-07c0-4617-9fb9-de876bed5e12): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:05:41 crc kubenswrapper[4934]: E1227 08:05:41.038139 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-api-5cc8cdc7b7-mmj42" podUID="a23b3091-07c0-4617-9fb9-de876bed5e12" Dec 27 08:05:41 crc kubenswrapper[4934]: I1227 08:05:41.169611 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6b7b667979-76g9s" podUID="ea0027c6-81c4-4d65-96a6-43eb035d2043" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.194:5353: connect: connection refused" Dec 27 08:05:41 crc kubenswrapper[4934]: I1227 08:05:41.343786 4934 scope.go:117] "RemoveContainer" containerID="f7c0182c802c5478d3e949a8b8cfad73ed9ed473d0b1b437a975df41dd930ac0" Dec 27 08:05:41 crc kubenswrapper[4934]: I1227 08:05:41.540646 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="327eb20e-e905-48a6-99b8-9209813e3330" path="/var/lib/kubelet/pods/327eb20e-e905-48a6-99b8-9209813e3330/volumes" Dec 27 08:05:41 crc kubenswrapper[4934]: I1227 08:05:41.651726 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-f7fc9c49c-vpz65" Dec 27 08:05:41 crc kubenswrapper[4934]: I1227 08:05:41.775956 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjc2p\" (UniqueName: \"kubernetes.io/projected/7b983753-0d99-48aa-82e9-c5a0643cf4c3-kube-api-access-wjc2p\") pod \"7b983753-0d99-48aa-82e9-c5a0643cf4c3\" (UID: \"7b983753-0d99-48aa-82e9-c5a0643cf4c3\") " Dec 27 08:05:41 crc kubenswrapper[4934]: I1227 08:05:41.776170 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b983753-0d99-48aa-82e9-c5a0643cf4c3-combined-ca-bundle\") pod \"7b983753-0d99-48aa-82e9-c5a0643cf4c3\" (UID: \"7b983753-0d99-48aa-82e9-c5a0643cf4c3\") " Dec 27 08:05:41 crc kubenswrapper[4934]: I1227 08:05:41.776390 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b983753-0d99-48aa-82e9-c5a0643cf4c3-config-data\") pod \"7b983753-0d99-48aa-82e9-c5a0643cf4c3\" (UID: \"7b983753-0d99-48aa-82e9-c5a0643cf4c3\") " Dec 27 08:05:41 crc kubenswrapper[4934]: I1227 08:05:41.776418 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b983753-0d99-48aa-82e9-c5a0643cf4c3-config-data-custom\") pod \"7b983753-0d99-48aa-82e9-c5a0643cf4c3\" (UID: \"7b983753-0d99-48aa-82e9-c5a0643cf4c3\") " Dec 27 08:05:41 crc kubenswrapper[4934]: I1227 08:05:41.787318 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b983753-0d99-48aa-82e9-c5a0643cf4c3-kube-api-access-wjc2p" (OuterVolumeSpecName: "kube-api-access-wjc2p") pod "7b983753-0d99-48aa-82e9-c5a0643cf4c3" (UID: "7b983753-0d99-48aa-82e9-c5a0643cf4c3"). InnerVolumeSpecName "kube-api-access-wjc2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:41 crc kubenswrapper[4934]: I1227 08:05:41.798306 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b983753-0d99-48aa-82e9-c5a0643cf4c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b983753-0d99-48aa-82e9-c5a0643cf4c3" (UID: "7b983753-0d99-48aa-82e9-c5a0643cf4c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:41 crc kubenswrapper[4934]: I1227 08:05:41.798354 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b983753-0d99-48aa-82e9-c5a0643cf4c3-config-data" (OuterVolumeSpecName: "config-data") pod "7b983753-0d99-48aa-82e9-c5a0643cf4c3" (UID: "7b983753-0d99-48aa-82e9-c5a0643cf4c3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:41 crc kubenswrapper[4934]: I1227 08:05:41.798459 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b983753-0d99-48aa-82e9-c5a0643cf4c3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7b983753-0d99-48aa-82e9-c5a0643cf4c3" (UID: "7b983753-0d99-48aa-82e9-c5a0643cf4c3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:41 crc kubenswrapper[4934]: I1227 08:05:41.897913 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b983753-0d99-48aa-82e9-c5a0643cf4c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:41 crc kubenswrapper[4934]: I1227 08:05:41.898261 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b983753-0d99-48aa-82e9-c5a0643cf4c3-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:41 crc kubenswrapper[4934]: I1227 08:05:41.898271 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7b983753-0d99-48aa-82e9-c5a0643cf4c3-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:41 crc kubenswrapper[4934]: I1227 08:05:41.898279 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjc2p\" (UniqueName: \"kubernetes.io/projected/7b983753-0d99-48aa-82e9-c5a0643cf4c3-kube-api-access-wjc2p\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:41 crc kubenswrapper[4934]: I1227 08:05:41.996691 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-f7fc9c49c-vpz65" event={"ID":"7b983753-0d99-48aa-82e9-c5a0643cf4c3","Type":"ContainerDied","Data":"07df07767df3d7d1ce37f9daa6340525c65a8f871cc7713828cf4aeee4432498"} Dec 27 08:05:41 crc kubenswrapper[4934]: I1227 08:05:41.996783 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-f7fc9c49c-vpz65" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.330646 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-f7fc9c49c-vpz65"] Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.342540 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-f7fc9c49c-vpz65"] Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.359435 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.363896 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76886c86d8-c6mdr" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.421472 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-combined-ca-bundle\") pod \"c67ffca3-2539-4344-848e-e251e564f42b\" (UID: \"c67ffca3-2539-4344-848e-e251e564f42b\") " Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.421761 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-dns-swift-storage-0\") pod \"ea0027c6-81c4-4d65-96a6-43eb035d2043\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.421864 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-config\") pod \"c67ffca3-2539-4344-848e-e251e564f42b\" (UID: \"c67ffca3-2539-4344-848e-e251e564f42b\") " Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.421937 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-config\") pod \"ea0027c6-81c4-4d65-96a6-43eb035d2043\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.422395 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-ovsdbserver-sb\") pod \"ea0027c6-81c4-4d65-96a6-43eb035d2043\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.422488 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-httpd-config\") pod \"c67ffca3-2539-4344-848e-e251e564f42b\" (UID: \"c67ffca3-2539-4344-848e-e251e564f42b\") " Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.422597 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m62pd\" (UniqueName: \"kubernetes.io/projected/ea0027c6-81c4-4d65-96a6-43eb035d2043-kube-api-access-m62pd\") pod \"ea0027c6-81c4-4d65-96a6-43eb035d2043\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.422672 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-ovndb-tls-certs\") pod \"c67ffca3-2539-4344-848e-e251e564f42b\" (UID: \"c67ffca3-2539-4344-848e-e251e564f42b\") " Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.422754 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8p5mf\" (UniqueName: \"kubernetes.io/projected/c67ffca3-2539-4344-848e-e251e564f42b-kube-api-access-8p5mf\") pod \"c67ffca3-2539-4344-848e-e251e564f42b\" (UID: \"c67ffca3-2539-4344-848e-e251e564f42b\") " Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.422819 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-dns-svc\") pod \"ea0027c6-81c4-4d65-96a6-43eb035d2043\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.422926 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-ovsdbserver-nb\") pod \"ea0027c6-81c4-4d65-96a6-43eb035d2043\" (UID: \"ea0027c6-81c4-4d65-96a6-43eb035d2043\") " Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.436074 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea0027c6-81c4-4d65-96a6-43eb035d2043-kube-api-access-m62pd" (OuterVolumeSpecName: "kube-api-access-m62pd") pod "ea0027c6-81c4-4d65-96a6-43eb035d2043" (UID: "ea0027c6-81c4-4d65-96a6-43eb035d2043"). InnerVolumeSpecName "kube-api-access-m62pd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.439045 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c67ffca3-2539-4344-848e-e251e564f42b-kube-api-access-8p5mf" (OuterVolumeSpecName: "kube-api-access-8p5mf") pod "c67ffca3-2539-4344-848e-e251e564f42b" (UID: "c67ffca3-2539-4344-848e-e251e564f42b"). InnerVolumeSpecName "kube-api-access-8p5mf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.448467 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "c67ffca3-2539-4344-848e-e251e564f42b" (UID: "c67ffca3-2539-4344-848e-e251e564f42b"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.489184 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-6bdc8c9885-7z5pz" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.533615 4934 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.533646 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m62pd\" (UniqueName: \"kubernetes.io/projected/ea0027c6-81c4-4d65-96a6-43eb035d2043-kube-api-access-m62pd\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.533656 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8p5mf\" (UniqueName: \"kubernetes.io/projected/c67ffca3-2539-4344-848e-e251e564f42b-kube-api-access-8p5mf\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.573714 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ea0027c6-81c4-4d65-96a6-43eb035d2043" (UID: "ea0027c6-81c4-4d65-96a6-43eb035d2043"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.583891 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c67ffca3-2539-4344-848e-e251e564f42b" (UID: "c67ffca3-2539-4344-848e-e251e564f42b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.584357 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ea0027c6-81c4-4d65-96a6-43eb035d2043" (UID: "ea0027c6-81c4-4d65-96a6-43eb035d2043"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.598415 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ea0027c6-81c4-4d65-96a6-43eb035d2043" (UID: "ea0027c6-81c4-4d65-96a6-43eb035d2043"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.628551 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-config" (OuterVolumeSpecName: "config") pod "c67ffca3-2539-4344-848e-e251e564f42b" (UID: "c67ffca3-2539-4344-848e-e251e564f42b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.637139 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.637166 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.637177 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.637186 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.637194 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.641354 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "c67ffca3-2539-4344-848e-e251e564f42b" (UID: "c67ffca3-2539-4344-848e-e251e564f42b"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.679062 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ea0027c6-81c4-4d65-96a6-43eb035d2043" (UID: "ea0027c6-81c4-4d65-96a6-43eb035d2043"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.681597 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-config" (OuterVolumeSpecName: "config") pod "ea0027c6-81c4-4d65-96a6-43eb035d2043" (UID: "ea0027c6-81c4-4d65-96a6-43eb035d2043"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.739444 4934 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.739474 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea0027c6-81c4-4d65-96a6-43eb035d2043-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.739484 4934 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c67ffca3-2539-4344-848e-e251e564f42b-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.836634 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-cf699bf96-qsv9c"] Dec 27 08:05:42 crc kubenswrapper[4934]: W1227 08:05:42.847295 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67b18dc3_52fb_4d6c_81de_ec66d3e19e57.slice/crio-4f7e9d0fc3bef95eb6260f173ed1572969592456b56d7f09f2545820b169fc35 WatchSource:0}: Error finding container 4f7e9d0fc3bef95eb6260f173ed1572969592456b56d7f09f2545820b169fc35: Status 404 returned error can't find the container with id 4f7e9d0fc3bef95eb6260f173ed1572969592456b56d7f09f2545820b169fc35 Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.849209 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-785dfb69f-wftff"] Dec 27 08:05:42 crc kubenswrapper[4934]: I1227 08:05:42.859874 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-7f59f88684-gxs5t"] Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.018980 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5cc8cdc7b7-mmj42" Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.038362 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-76g9s" event={"ID":"ea0027c6-81c4-4d65-96a6-43eb035d2043","Type":"ContainerDied","Data":"2897983f242ebc3009030fc753cbef7e0671608abac9db84c943b84ea43d670b"} Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.038439 4934 scope.go:117] "RemoveContainer" containerID="416ae62c615270f259b28149ef9568a4f45c93acc55599fbdfd7470a080632e2" Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.038666 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-76g9s" Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.043917 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5cc8cdc7b7-mmj42" event={"ID":"a23b3091-07c0-4617-9fb9-de876bed5e12","Type":"ContainerDied","Data":"df28844f078156278c57e2df6e32ed3ebf77c1d7b8ad2afa7d23b4993f155d80"} Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.044023 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5cc8cdc7b7-mmj42" Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.044749 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a23b3091-07c0-4617-9fb9-de876bed5e12-config-data\") pod \"a23b3091-07c0-4617-9fb9-de876bed5e12\" (UID: \"a23b3091-07c0-4617-9fb9-de876bed5e12\") " Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.044823 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdxpj\" (UniqueName: \"kubernetes.io/projected/a23b3091-07c0-4617-9fb9-de876bed5e12-kube-api-access-pdxpj\") pod \"a23b3091-07c0-4617-9fb9-de876bed5e12\" (UID: \"a23b3091-07c0-4617-9fb9-de876bed5e12\") " Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.044918 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a23b3091-07c0-4617-9fb9-de876bed5e12-combined-ca-bundle\") pod \"a23b3091-07c0-4617-9fb9-de876bed5e12\" (UID: \"a23b3091-07c0-4617-9fb9-de876bed5e12\") " Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.045050 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a23b3091-07c0-4617-9fb9-de876bed5e12-config-data-custom\") pod \"a23b3091-07c0-4617-9fb9-de876bed5e12\" (UID: \"a23b3091-07c0-4617-9fb9-de876bed5e12\") " Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.055765 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a23b3091-07c0-4617-9fb9-de876bed5e12-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a23b3091-07c0-4617-9fb9-de876bed5e12" (UID: "a23b3091-07c0-4617-9fb9-de876bed5e12"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.062435 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a23b3091-07c0-4617-9fb9-de876bed5e12-config-data" (OuterVolumeSpecName: "config-data") pod "a23b3091-07c0-4617-9fb9-de876bed5e12" (UID: "a23b3091-07c0-4617-9fb9-de876bed5e12"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.055995 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-785dfb69f-wftff" event={"ID":"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b","Type":"ContainerStarted","Data":"0599de68b5101a784656b736a194b29dde57feeed87b0890540c39580210147f"} Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.065642 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a23b3091-07c0-4617-9fb9-de876bed5e12-kube-api-access-pdxpj" (OuterVolumeSpecName: "kube-api-access-pdxpj") pod "a23b3091-07c0-4617-9fb9-de876bed5e12" (UID: "a23b3091-07c0-4617-9fb9-de876bed5e12"). InnerVolumeSpecName "kube-api-access-pdxpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.085280 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-db89fbccf-w9xvz"] Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.085929 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a23b3091-07c0-4617-9fb9-de876bed5e12-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a23b3091-07c0-4617-9fb9-de876bed5e12" (UID: "a23b3091-07c0-4617-9fb9-de876bed5e12"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.085969 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76886c86d8-c6mdr" event={"ID":"c67ffca3-2539-4344-848e-e251e564f42b","Type":"ContainerDied","Data":"3fc3110b2306014fade5fc8825927ac431b1dbf9b36adf451772a256afe222c4"} Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.085997 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76886c86d8-c6mdr" Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.088031 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deefee78-9e75-44e2-b72a-08aeb8b97ac4","Type":"ContainerStarted","Data":"c7aef981d05c3a0478c37d7d09c12e1d2edce81b6b397585f3f3259f03f9b58e"} Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.091430 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-cf699bf96-qsv9c" event={"ID":"67b18dc3-52fb-4d6c-81de-ec66d3e19e57","Type":"ContainerStarted","Data":"4f7e9d0fc3bef95eb6260f173ed1572969592456b56d7f09f2545820b169fc35"} Dec 27 08:05:43 crc kubenswrapper[4934]: W1227 08:05:43.099106 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1515c241_e28d_49d7_a3f3_790aa082251c.slice/crio-6d6cdc0c6658bbfb32e7990958b1a7e92df2cc9548f50d710ece359edd8cf5a4 WatchSource:0}: Error finding container 6d6cdc0c6658bbfb32e7990958b1a7e92df2cc9548f50d710ece359edd8cf5a4: Status 404 returned error can't find the container with id 6d6cdc0c6658bbfb32e7990958b1a7e92df2cc9548f50d710ece359edd8cf5a4 Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.102262 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7f59f88684-gxs5t" event={"ID":"1182d8a5-1924-45e8-b13f-18fc3b25cf9c","Type":"ContainerStarted","Data":"56dc1ca7f2bc570dc0724c3da2a278079753d432d38acf20a66b547d90937515"} Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.110550 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c","Type":"ContainerStarted","Data":"cf4a8a25e30cc4b825c310997855c40903649e7ba7f340b290f00f224f9f8056"} Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.128261 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6f98b67c6-55vgd"] Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.149936 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a23b3091-07c0-4617-9fb9-de876bed5e12-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.149968 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a23b3091-07c0-4617-9fb9-de876bed5e12-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.149981 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdxpj\" (UniqueName: \"kubernetes.io/projected/a23b3091-07c0-4617-9fb9-de876bed5e12-kube-api-access-pdxpj\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.149991 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a23b3091-07c0-4617-9fb9-de876bed5e12-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.203451 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-76886c86d8-c6mdr"] Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.217333 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-76886c86d8-c6mdr"] Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.225588 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-76g9s"] Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.234884 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-76g9s"] Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.424108 4934 scope.go:117] "RemoveContainer" containerID="3dd317627a8b1981085d0b2a2bd00372273d1ad0d839ba67ada22bbbb49ef0f0" Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.427936 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5cc8cdc7b7-mmj42"] Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.445586 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-5cc8cdc7b7-mmj42"] Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.494658 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b983753-0d99-48aa-82e9-c5a0643cf4c3" path="/var/lib/kubelet/pods/7b983753-0d99-48aa-82e9-c5a0643cf4c3/volumes" Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.495166 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a23b3091-07c0-4617-9fb9-de876bed5e12" path="/var/lib/kubelet/pods/a23b3091-07c0-4617-9fb9-de876bed5e12/volumes" Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.495543 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c67ffca3-2539-4344-848e-e251e564f42b" path="/var/lib/kubelet/pods/c67ffca3-2539-4344-848e-e251e564f42b/volumes" Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.496164 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea0027c6-81c4-4d65-96a6-43eb035d2043" path="/var/lib/kubelet/pods/ea0027c6-81c4-4d65-96a6-43eb035d2043/volumes" Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.566888 4934 scope.go:117] "RemoveContainer" containerID="67e0ab891a2e8f56faeae88e28a6fca9233b0b966fcab3583810609696d05004" Dec 27 08:05:43 crc kubenswrapper[4934]: I1227 08:05:43.604068 4934 scope.go:117] "RemoveContainer" containerID="267c584dc50d95618fa637ec47ff3c93f9e2ea0d158207d0591b2838c4afa90d" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.131124 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6f98b67c6-55vgd" event={"ID":"1515c241-e28d-49d7-a3f3-790aa082251c","Type":"ContainerStarted","Data":"80405edd7bb455015027bd7125e6267a53d052b855fe164cffc53d0ef0acd2da"} Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.131472 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6f98b67c6-55vgd" event={"ID":"1515c241-e28d-49d7-a3f3-790aa082251c","Type":"ContainerStarted","Data":"6d6cdc0c6658bbfb32e7990958b1a7e92df2cc9548f50d710ece359edd8cf5a4"} Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.131502 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-6f98b67c6-55vgd" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.133816 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deefee78-9e75-44e2-b72a-08aeb8b97ac4","Type":"ContainerStarted","Data":"ff283c40ff7fc3fffc16d3b5341ce187559ba24fdae536c0f89834ef0197cc75"} Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.135034 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-db89fbccf-w9xvz" event={"ID":"7237f7ae-a705-41af-ace6-27bafbba974d","Type":"ContainerStarted","Data":"b7028b9d51c28b6cc058a2a2997853affa85d43c9ef8e0adc5df2b05a2058071"} Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.141536 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c","Type":"ContainerStarted","Data":"a1d593c5cf64cc6fe0db885ff877caf0fcbaafeaf7c140c510a4d56580261bc5"} Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.158542 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-6f98b67c6-55vgd" podStartSLOduration=12.158527105 podStartE2EDuration="12.158527105s" podCreationTimestamp="2025-12-27 08:05:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:05:44.152986348 +0000 UTC m=+1404.973426972" watchObservedRunningTime="2025-12-27 08:05:44.158527105 +0000 UTC m=+1404.978967699" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.174916 4934 generic.go:334] "Generic (PLEG): container finished" podID="f994d21a-55df-4ea2-a86d-41d18d10d4d1" containerID="d16fe7c3850179dc183ac942ae64cb6065521e25fa98f3bc7089f111a9b8c0a1" exitCode=0 Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.175174 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f994d21a-55df-4ea2-a86d-41d18d10d4d1","Type":"ContainerDied","Data":"d16fe7c3850179dc183ac942ae64cb6065521e25fa98f3bc7089f111a9b8c0a1"} Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.179879 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=7.584379522 podStartE2EDuration="23.17986117s" podCreationTimestamp="2025-12-27 08:05:21 +0000 UTC" firstStartedPulling="2025-12-27 08:05:24.146637023 +0000 UTC m=+1384.967077617" lastFinishedPulling="2025-12-27 08:05:39.742118671 +0000 UTC m=+1400.562559265" observedRunningTime="2025-12-27 08:05:44.16968956 +0000 UTC m=+1404.990130154" watchObservedRunningTime="2025-12-27 08:05:44.17986117 +0000 UTC m=+1405.000301764" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.500841 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.593447 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f994d21a-55df-4ea2-a86d-41d18d10d4d1-logs\") pod \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.593546 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-config-data\") pod \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.593762 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-combined-ca-bundle\") pod \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.593880 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\") pod \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.593912 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfv77\" (UniqueName: \"kubernetes.io/projected/f994d21a-55df-4ea2-a86d-41d18d10d4d1-kube-api-access-xfv77\") pod \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.593934 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f994d21a-55df-4ea2-a86d-41d18d10d4d1-httpd-run\") pod \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.593965 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-scripts\") pod \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.594038 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-public-tls-certs\") pod \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\" (UID: \"f994d21a-55df-4ea2-a86d-41d18d10d4d1\") " Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.594179 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f994d21a-55df-4ea2-a86d-41d18d10d4d1-logs" (OuterVolumeSpecName: "logs") pod "f994d21a-55df-4ea2-a86d-41d18d10d4d1" (UID: "f994d21a-55df-4ea2-a86d-41d18d10d4d1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.594570 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f994d21a-55df-4ea2-a86d-41d18d10d4d1-logs\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.595423 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f994d21a-55df-4ea2-a86d-41d18d10d4d1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f994d21a-55df-4ea2-a86d-41d18d10d4d1" (UID: "f994d21a-55df-4ea2-a86d-41d18d10d4d1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.601913 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f994d21a-55df-4ea2-a86d-41d18d10d4d1-kube-api-access-xfv77" (OuterVolumeSpecName: "kube-api-access-xfv77") pod "f994d21a-55df-4ea2-a86d-41d18d10d4d1" (UID: "f994d21a-55df-4ea2-a86d-41d18d10d4d1"). InnerVolumeSpecName "kube-api-access-xfv77". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.607482 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-scripts" (OuterVolumeSpecName: "scripts") pod "f994d21a-55df-4ea2-a86d-41d18d10d4d1" (UID: "f994d21a-55df-4ea2-a86d-41d18d10d4d1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.642278 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6" (OuterVolumeSpecName: "glance") pod "f994d21a-55df-4ea2-a86d-41d18d10d4d1" (UID: "f994d21a-55df-4ea2-a86d-41d18d10d4d1"). InnerVolumeSpecName "pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.698572 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\") on node \"crc\" " Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.698603 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfv77\" (UniqueName: \"kubernetes.io/projected/f994d21a-55df-4ea2-a86d-41d18d10d4d1-kube-api-access-xfv77\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.698614 4934 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f994d21a-55df-4ea2-a86d-41d18d10d4d1-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.698626 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.821584 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-zkwng"] Dec 27 08:05:44 crc kubenswrapper[4934]: E1227 08:05:44.822367 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="327eb20e-e905-48a6-99b8-9209813e3330" containerName="barbican-api-log" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.822382 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="327eb20e-e905-48a6-99b8-9209813e3330" containerName="barbican-api-log" Dec 27 08:05:44 crc kubenswrapper[4934]: E1227 08:05:44.822400 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea0027c6-81c4-4d65-96a6-43eb035d2043" containerName="init" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.822406 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea0027c6-81c4-4d65-96a6-43eb035d2043" containerName="init" Dec 27 08:05:44 crc kubenswrapper[4934]: E1227 08:05:44.822414 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f994d21a-55df-4ea2-a86d-41d18d10d4d1" containerName="glance-httpd" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.822420 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f994d21a-55df-4ea2-a86d-41d18d10d4d1" containerName="glance-httpd" Dec 27 08:05:44 crc kubenswrapper[4934]: E1227 08:05:44.822431 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c67ffca3-2539-4344-848e-e251e564f42b" containerName="neutron-httpd" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.822437 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c67ffca3-2539-4344-848e-e251e564f42b" containerName="neutron-httpd" Dec 27 08:05:44 crc kubenswrapper[4934]: E1227 08:05:44.822448 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="327eb20e-e905-48a6-99b8-9209813e3330" containerName="barbican-api" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.822454 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="327eb20e-e905-48a6-99b8-9209813e3330" containerName="barbican-api" Dec 27 08:05:44 crc kubenswrapper[4934]: E1227 08:05:44.822468 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea0027c6-81c4-4d65-96a6-43eb035d2043" containerName="dnsmasq-dns" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.822474 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea0027c6-81c4-4d65-96a6-43eb035d2043" containerName="dnsmasq-dns" Dec 27 08:05:44 crc kubenswrapper[4934]: E1227 08:05:44.822508 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f994d21a-55df-4ea2-a86d-41d18d10d4d1" containerName="glance-log" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.822518 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f994d21a-55df-4ea2-a86d-41d18d10d4d1" containerName="glance-log" Dec 27 08:05:44 crc kubenswrapper[4934]: E1227 08:05:44.822546 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c67ffca3-2539-4344-848e-e251e564f42b" containerName="neutron-api" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.822551 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c67ffca3-2539-4344-848e-e251e564f42b" containerName="neutron-api" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.822763 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f994d21a-55df-4ea2-a86d-41d18d10d4d1" containerName="glance-log" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.822776 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="327eb20e-e905-48a6-99b8-9209813e3330" containerName="barbican-api-log" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.822787 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c67ffca3-2539-4344-848e-e251e564f42b" containerName="neutron-api" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.822795 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c67ffca3-2539-4344-848e-e251e564f42b" containerName="neutron-httpd" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.822802 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="327eb20e-e905-48a6-99b8-9209813e3330" containerName="barbican-api" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.822819 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea0027c6-81c4-4d65-96a6-43eb035d2043" containerName="dnsmasq-dns" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.822827 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f994d21a-55df-4ea2-a86d-41d18d10d4d1" containerName="glance-httpd" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.823736 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-zkwng" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.859122 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-zkwng"] Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.865483 4934 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.865638 4934 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6") on node "crc" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.906952 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz8hp\" (UniqueName: \"kubernetes.io/projected/acbab7d1-71c2-4bf6-888b-b0e45005d374-kube-api-access-tz8hp\") pod \"nova-api-db-create-zkwng\" (UID: \"acbab7d1-71c2-4bf6-888b-b0e45005d374\") " pod="openstack/nova-api-db-create-zkwng" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.907068 4934 reconciler_common.go:293] "Volume detached for volume \"pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.913473 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f994d21a-55df-4ea2-a86d-41d18d10d4d1" (UID: "f994d21a-55df-4ea2-a86d-41d18d10d4d1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.929597 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-6kjfr"] Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.937406 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6kjfr" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.984237 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f994d21a-55df-4ea2-a86d-41d18d10d4d1" (UID: "f994d21a-55df-4ea2-a86d-41d18d10d4d1"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:44 crc kubenswrapper[4934]: I1227 08:05:44.984312 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-6kjfr"] Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.008697 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz8hp\" (UniqueName: \"kubernetes.io/projected/acbab7d1-71c2-4bf6-888b-b0e45005d374-kube-api-access-tz8hp\") pod \"nova-api-db-create-zkwng\" (UID: \"acbab7d1-71c2-4bf6-888b-b0e45005d374\") " pod="openstack/nova-api-db-create-zkwng" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.008880 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.008895 4934 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.023881 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-config-data" (OuterVolumeSpecName: "config-data") pod "f994d21a-55df-4ea2-a86d-41d18d10d4d1" (UID: "f994d21a-55df-4ea2-a86d-41d18d10d4d1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.028783 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-xbmch"] Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.039640 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xbmch" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.042944 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz8hp\" (UniqueName: \"kubernetes.io/projected/acbab7d1-71c2-4bf6-888b-b0e45005d374-kube-api-access-tz8hp\") pod \"nova-api-db-create-zkwng\" (UID: \"acbab7d1-71c2-4bf6-888b-b0e45005d374\") " pod="openstack/nova-api-db-create-zkwng" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.055950 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xbmch"] Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.111233 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvgfc\" (UniqueName: \"kubernetes.io/projected/8e8b0569-1879-41e6-a91c-7e1b812ff6d5-kube-api-access-gvgfc\") pod \"nova-cell0-db-create-6kjfr\" (UID: \"8e8b0569-1879-41e6-a91c-7e1b812ff6d5\") " pod="openstack/nova-cell0-db-create-6kjfr" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.112268 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f994d21a-55df-4ea2-a86d-41d18d10d4d1-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.188015 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f994d21a-55df-4ea2-a86d-41d18d10d4d1","Type":"ContainerDied","Data":"5aa76dc8fd33bf617d82ad0e757ea43d3b9145049183265764b584931927502a"} Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.188042 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.188111 4934 scope.go:117] "RemoveContainer" containerID="d16fe7c3850179dc183ac942ae64cb6065521e25fa98f3bc7089f111a9b8c0a1" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.190019 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-785dfb69f-wftff" event={"ID":"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b","Type":"ContainerStarted","Data":"d8ca97c2fc969b024a50a5e3cb3016115c652ed613bdfbefeb1b6bbba9809849"} Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.196617 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-zkwng" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.198383 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7f59f88684-gxs5t" event={"ID":"1182d8a5-1924-45e8-b13f-18fc3b25cf9c","Type":"ContainerStarted","Data":"23659a47b729f36f776ec2474b2cfadbe9a6c80b4857f65417410ced6b9c3a5f"} Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.198633 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7f59f88684-gxs5t" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.214307 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc6fr\" (UniqueName: \"kubernetes.io/projected/5cbaa946-14d7-4839-9541-772b8c5755c0-kube-api-access-wc6fr\") pod \"nova-cell1-db-create-xbmch\" (UID: \"5cbaa946-14d7-4839-9541-772b8c5755c0\") " pod="openstack/nova-cell1-db-create-xbmch" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.214419 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvgfc\" (UniqueName: \"kubernetes.io/projected/8e8b0569-1879-41e6-a91c-7e1b812ff6d5-kube-api-access-gvgfc\") pod \"nova-cell0-db-create-6kjfr\" (UID: \"8e8b0569-1879-41e6-a91c-7e1b812ff6d5\") " pod="openstack/nova-cell0-db-create-6kjfr" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.218780 4934 scope.go:117] "RemoveContainer" containerID="5738140d2bba76b7bc6a908b15245ca158c60b158eba7d68b51d42af6c3a07b2" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.242725 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvgfc\" (UniqueName: \"kubernetes.io/projected/8e8b0569-1879-41e6-a91c-7e1b812ff6d5-kube-api-access-gvgfc\") pod \"nova-cell0-db-create-6kjfr\" (UID: \"8e8b0569-1879-41e6-a91c-7e1b812ff6d5\") " pod="openstack/nova-cell0-db-create-6kjfr" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.293433 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-7f59f88684-gxs5t" podStartSLOduration=12.241000788 podStartE2EDuration="13.293415162s" podCreationTimestamp="2025-12-27 08:05:32 +0000 UTC" firstStartedPulling="2025-12-27 08:05:42.917906861 +0000 UTC m=+1403.738347455" lastFinishedPulling="2025-12-27 08:05:43.970321235 +0000 UTC m=+1404.790761829" observedRunningTime="2025-12-27 08:05:45.22638982 +0000 UTC m=+1406.046830414" watchObservedRunningTime="2025-12-27 08:05:45.293415162 +0000 UTC m=+1406.113855756" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.295555 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.320386 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc6fr\" (UniqueName: \"kubernetes.io/projected/5cbaa946-14d7-4839-9541-772b8c5755c0-kube-api-access-wc6fr\") pod \"nova-cell1-db-create-xbmch\" (UID: \"5cbaa946-14d7-4839-9541-772b8c5755c0\") " pod="openstack/nova-cell1-db-create-xbmch" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.323749 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.334492 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.336595 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.341360 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.341625 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.343352 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.370379 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc6fr\" (UniqueName: \"kubernetes.io/projected/5cbaa946-14d7-4839-9541-772b8c5755c0-kube-api-access-wc6fr\") pod \"nova-cell1-db-create-xbmch\" (UID: \"5cbaa946-14d7-4839-9541-772b8c5755c0\") " pod="openstack/nova-cell1-db-create-xbmch" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.412474 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6kjfr" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.489953 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f994d21a-55df-4ea2-a86d-41d18d10d4d1" path="/var/lib/kubelet/pods/f994d21a-55df-4ea2-a86d-41d18d10d4d1/volumes" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.528487 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3152d244-2397-469d-a570-d4fd09ee9a0d-logs\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.528684 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7sjz\" (UniqueName: \"kubernetes.io/projected/3152d244-2397-469d-a570-d4fd09ee9a0d-kube-api-access-b7sjz\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.528740 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3152d244-2397-469d-a570-d4fd09ee9a0d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.528874 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3152d244-2397-469d-a570-d4fd09ee9a0d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.528906 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3152d244-2397-469d-a570-d4fd09ee9a0d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.528935 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3152d244-2397-469d-a570-d4fd09ee9a0d-scripts\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.529028 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3152d244-2397-469d-a570-d4fd09ee9a0d-config-data\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.529074 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.544898 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xbmch" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.630958 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3152d244-2397-469d-a570-d4fd09ee9a0d-config-data\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.631012 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.631160 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3152d244-2397-469d-a570-d4fd09ee9a0d-logs\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.631247 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7sjz\" (UniqueName: \"kubernetes.io/projected/3152d244-2397-469d-a570-d4fd09ee9a0d-kube-api-access-b7sjz\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.631270 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3152d244-2397-469d-a570-d4fd09ee9a0d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.631320 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3152d244-2397-469d-a570-d4fd09ee9a0d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.631341 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3152d244-2397-469d-a570-d4fd09ee9a0d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.631358 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3152d244-2397-469d-a570-d4fd09ee9a0d-scripts\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.636242 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3152d244-2397-469d-a570-d4fd09ee9a0d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.638381 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3152d244-2397-469d-a570-d4fd09ee9a0d-logs\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.649477 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3152d244-2397-469d-a570-d4fd09ee9a0d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.651736 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.651847 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9f50bff41a79b6548a20cce2c4162bc1d0a0a91c484e04e21b6d749439a0f41c/globalmount\"" pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.652756 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3152d244-2397-469d-a570-d4fd09ee9a0d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.653499 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7sjz\" (UniqueName: \"kubernetes.io/projected/3152d244-2397-469d-a570-d4fd09ee9a0d-kube-api-access-b7sjz\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.655991 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3152d244-2397-469d-a570-d4fd09ee9a0d-scripts\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.658145 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3152d244-2397-469d-a570-d4fd09ee9a0d-config-data\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.723525 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-de2595de-29dc-46f5-b0cc-a82acce6eab6\") pod \"glance-default-external-api-0\" (UID: \"3152d244-2397-469d-a570-d4fd09ee9a0d\") " pod="openstack/glance-default-external-api-0" Dec 27 08:05:45 crc kubenswrapper[4934]: I1227 08:05:45.744765 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 27 08:05:46 crc kubenswrapper[4934]: I1227 08:05:46.050764 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-zkwng"] Dec 27 08:05:46 crc kubenswrapper[4934]: I1227 08:05:46.128890 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-6kjfr"] Dec 27 08:05:46 crc kubenswrapper[4934]: I1227 08:05:46.241193 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-zkwng" event={"ID":"acbab7d1-71c2-4bf6-888b-b0e45005d374","Type":"ContainerStarted","Data":"50780ad4e074d285e8be33becca6d152a58c0f23c892747efba4785855a0da0a"} Dec 27 08:05:46 crc kubenswrapper[4934]: I1227 08:05:46.259868 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deefee78-9e75-44e2-b72a-08aeb8b97ac4","Type":"ContainerStarted","Data":"f6b25f19435e4558d201238c1e37486a77141d8977bbfd3f0a3dfb2249350764"} Dec 27 08:05:46 crc kubenswrapper[4934]: I1227 08:05:46.267213 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-cf699bf96-qsv9c" event={"ID":"67b18dc3-52fb-4d6c-81de-ec66d3e19e57","Type":"ContainerStarted","Data":"98466e60fc6c1c49235813b1ce6248fb7e70cfdf95fa0e64f6a8ccdac3d08039"} Dec 27 08:05:46 crc kubenswrapper[4934]: I1227 08:05:46.267316 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:46 crc kubenswrapper[4934]: I1227 08:05:46.270137 4934 generic.go:334] "Generic (PLEG): container finished" podID="1182d8a5-1924-45e8-b13f-18fc3b25cf9c" containerID="23659a47b729f36f776ec2474b2cfadbe9a6c80b4857f65417410ced6b9c3a5f" exitCode=1 Dec 27 08:05:46 crc kubenswrapper[4934]: I1227 08:05:46.270201 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7f59f88684-gxs5t" event={"ID":"1182d8a5-1924-45e8-b13f-18fc3b25cf9c","Type":"ContainerDied","Data":"23659a47b729f36f776ec2474b2cfadbe9a6c80b4857f65417410ced6b9c3a5f"} Dec 27 08:05:46 crc kubenswrapper[4934]: I1227 08:05:46.270997 4934 scope.go:117] "RemoveContainer" containerID="23659a47b729f36f776ec2474b2cfadbe9a6c80b4857f65417410ced6b9c3a5f" Dec 27 08:05:46 crc kubenswrapper[4934]: I1227 08:05:46.283676 4934 generic.go:334] "Generic (PLEG): container finished" podID="7237f7ae-a705-41af-ace6-27bafbba974d" containerID="1377ab23df0dece01efb4cf0d4f7407e3556cc72553b80348dfa5f01d091d2f7" exitCode=1 Dec 27 08:05:46 crc kubenswrapper[4934]: I1227 08:05:46.283743 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-db89fbccf-w9xvz" event={"ID":"7237f7ae-a705-41af-ace6-27bafbba974d","Type":"ContainerDied","Data":"1377ab23df0dece01efb4cf0d4f7407e3556cc72553b80348dfa5f01d091d2f7"} Dec 27 08:05:46 crc kubenswrapper[4934]: I1227 08:05:46.284441 4934 scope.go:117] "RemoveContainer" containerID="1377ab23df0dece01efb4cf0d4f7407e3556cc72553b80348dfa5f01d091d2f7" Dec 27 08:05:46 crc kubenswrapper[4934]: I1227 08:05:46.296418 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6kjfr" event={"ID":"8e8b0569-1879-41e6-a91c-7e1b812ff6d5","Type":"ContainerStarted","Data":"25df38bd251d365dc60d919e45b39b0af84914b7d6987dd0a7132befd03cc81b"} Dec 27 08:05:46 crc kubenswrapper[4934]: I1227 08:05:46.296456 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xbmch"] Dec 27 08:05:46 crc kubenswrapper[4934]: I1227 08:05:46.296724 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:46 crc kubenswrapper[4934]: I1227 08:05:46.318218 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-cf699bf96-qsv9c" podStartSLOduration=9.88117019 podStartE2EDuration="11.318193415s" podCreationTimestamp="2025-12-27 08:05:35 +0000 UTC" firstStartedPulling="2025-12-27 08:05:42.894618047 +0000 UTC m=+1403.715058631" lastFinishedPulling="2025-12-27 08:05:44.331641262 +0000 UTC m=+1405.152081856" observedRunningTime="2025-12-27 08:05:46.28917814 +0000 UTC m=+1407.109618744" watchObservedRunningTime="2025-12-27 08:05:46.318193415 +0000 UTC m=+1407.138634009" Dec 27 08:05:46 crc kubenswrapper[4934]: I1227 08:05:46.367186 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-785dfb69f-wftff" podStartSLOduration=9.941313282 podStartE2EDuration="11.367160532s" podCreationTimestamp="2025-12-27 08:05:35 +0000 UTC" firstStartedPulling="2025-12-27 08:05:42.894941645 +0000 UTC m=+1403.715382239" lastFinishedPulling="2025-12-27 08:05:44.320788895 +0000 UTC m=+1405.141229489" observedRunningTime="2025-12-27 08:05:46.355927765 +0000 UTC m=+1407.176368359" watchObservedRunningTime="2025-12-27 08:05:46.367160532 +0000 UTC m=+1407.187601126" Dec 27 08:05:46 crc kubenswrapper[4934]: I1227 08:05:46.567684 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 27 08:05:46 crc kubenswrapper[4934]: I1227 08:05:46.589274 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.316399 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deefee78-9e75-44e2-b72a-08aeb8b97ac4","Type":"ContainerStarted","Data":"62d0acb0da855b7313ed569eab92f4246077da92cd6e3596cc95bd0942f77b98"} Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.317112 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="deefee78-9e75-44e2-b72a-08aeb8b97ac4" containerName="ceilometer-central-agent" containerID="cri-o://c7aef981d05c3a0478c37d7d09c12e1d2edce81b6b397585f3f3259f03f9b58e" gracePeriod=30 Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.317410 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.317724 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="deefee78-9e75-44e2-b72a-08aeb8b97ac4" containerName="proxy-httpd" containerID="cri-o://62d0acb0da855b7313ed569eab92f4246077da92cd6e3596cc95bd0942f77b98" gracePeriod=30 Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.317769 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="deefee78-9e75-44e2-b72a-08aeb8b97ac4" containerName="sg-core" containerID="cri-o://f6b25f19435e4558d201238c1e37486a77141d8977bbfd3f0a3dfb2249350764" gracePeriod=30 Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.317803 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="deefee78-9e75-44e2-b72a-08aeb8b97ac4" containerName="ceilometer-notification-agent" containerID="cri-o://ff283c40ff7fc3fffc16d3b5341ce187559ba24fdae536c0f89834ef0197cc75" gracePeriod=30 Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.341549 4934 generic.go:334] "Generic (PLEG): container finished" podID="1182d8a5-1924-45e8-b13f-18fc3b25cf9c" containerID="661d957736179cd056778b9ec417c69d5afd27f25d079ccc3e962bd155cf48d4" exitCode=1 Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.341854 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7f59f88684-gxs5t" event={"ID":"1182d8a5-1924-45e8-b13f-18fc3b25cf9c","Type":"ContainerDied","Data":"661d957736179cd056778b9ec417c69d5afd27f25d079ccc3e962bd155cf48d4"} Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.341938 4934 scope.go:117] "RemoveContainer" containerID="23659a47b729f36f776ec2474b2cfadbe9a6c80b4857f65417410ced6b9c3a5f" Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.342726 4934 scope.go:117] "RemoveContainer" containerID="661d957736179cd056778b9ec417c69d5afd27f25d079ccc3e962bd155cf48d4" Dec 27 08:05:47 crc kubenswrapper[4934]: E1227 08:05:47.343010 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-7f59f88684-gxs5t_openstack(1182d8a5-1924-45e8-b13f-18fc3b25cf9c)\"" pod="openstack/heat-cfnapi-7f59f88684-gxs5t" podUID="1182d8a5-1924-45e8-b13f-18fc3b25cf9c" Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.350457 4934 generic.go:334] "Generic (PLEG): container finished" podID="5cbaa946-14d7-4839-9541-772b8c5755c0" containerID="9559761b33b407ededc4d3d58c8f1366c61c5d156b946dc7bafd3405ad9b4c9b" exitCode=0 Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.350517 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xbmch" event={"ID":"5cbaa946-14d7-4839-9541-772b8c5755c0","Type":"ContainerDied","Data":"9559761b33b407ededc4d3d58c8f1366c61c5d156b946dc7bafd3405ad9b4c9b"} Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.350545 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xbmch" event={"ID":"5cbaa946-14d7-4839-9541-772b8c5755c0","Type":"ContainerStarted","Data":"7918ab665d6b8f1a34e6b900208cde6cdd31943a132df593a2560a44c146a91c"} Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.361004 4934 generic.go:334] "Generic (PLEG): container finished" podID="7237f7ae-a705-41af-ace6-27bafbba974d" containerID="fe3d37707fe9adabcbc2464fb446cba67915cd60ffb4fc6e35846c1b562835d2" exitCode=1 Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.361073 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-db89fbccf-w9xvz" event={"ID":"7237f7ae-a705-41af-ace6-27bafbba974d","Type":"ContainerDied","Data":"fe3d37707fe9adabcbc2464fb446cba67915cd60ffb4fc6e35846c1b562835d2"} Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.361932 4934 scope.go:117] "RemoveContainer" containerID="fe3d37707fe9adabcbc2464fb446cba67915cd60ffb4fc6e35846c1b562835d2" Dec 27 08:05:47 crc kubenswrapper[4934]: E1227 08:05:47.362205 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-db89fbccf-w9xvz_openstack(7237f7ae-a705-41af-ace6-27bafbba974d)\"" pod="openstack/heat-api-db89fbccf-w9xvz" podUID="7237f7ae-a705-41af-ace6-27bafbba974d" Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.367670 4934 generic.go:334] "Generic (PLEG): container finished" podID="8e8b0569-1879-41e6-a91c-7e1b812ff6d5" containerID="198266d3b59d65fb02b064c61f2f6e47c2b54fd06aa20323c6b09a9b554e7050" exitCode=0 Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.367752 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6kjfr" event={"ID":"8e8b0569-1879-41e6-a91c-7e1b812ff6d5","Type":"ContainerDied","Data":"198266d3b59d65fb02b064c61f2f6e47c2b54fd06aa20323c6b09a9b554e7050"} Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.382472 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.491355613 podStartE2EDuration="23.38245339s" podCreationTimestamp="2025-12-27 08:05:24 +0000 UTC" firstStartedPulling="2025-12-27 08:05:25.938578557 +0000 UTC m=+1386.759019151" lastFinishedPulling="2025-12-27 08:05:46.829676334 +0000 UTC m=+1407.650116928" observedRunningTime="2025-12-27 08:05:47.336984719 +0000 UTC m=+1408.157425313" watchObservedRunningTime="2025-12-27 08:05:47.38245339 +0000 UTC m=+1408.202893984" Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.384371 4934 generic.go:334] "Generic (PLEG): container finished" podID="acbab7d1-71c2-4bf6-888b-b0e45005d374" containerID="9e228948ddcf016bf0a216af729f9f9a1f35ee7a0eb7def58d6a6cff43b65cd3" exitCode=0 Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.384534 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-zkwng" event={"ID":"acbab7d1-71c2-4bf6-888b-b0e45005d374","Type":"ContainerDied","Data":"9e228948ddcf016bf0a216af729f9f9a1f35ee7a0eb7def58d6a6cff43b65cd3"} Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.397692 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3152d244-2397-469d-a570-d4fd09ee9a0d","Type":"ContainerStarted","Data":"379b47a5c5cecb5beeeccc6242ccaffeb40a8ddfdab5fcd6ddde18f7189275fc"} Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.397734 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3152d244-2397-469d-a570-d4fd09ee9a0d","Type":"ContainerStarted","Data":"bfc34d7a85b855538b54010e2d55ac5564cc0e09c5e142e016d6445c1e8bd67e"} Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.611142 4934 scope.go:117] "RemoveContainer" containerID="1377ab23df0dece01efb4cf0d4f7407e3556cc72553b80348dfa5f01d091d2f7" Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.791560 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-7f59f88684-gxs5t" Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.791606 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-7f59f88684-gxs5t" Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.836036 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-db89fbccf-w9xvz" Dec 27 08:05:47 crc kubenswrapper[4934]: I1227 08:05:47.836074 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-db89fbccf-w9xvz" Dec 27 08:05:48 crc kubenswrapper[4934]: I1227 08:05:48.409312 4934 scope.go:117] "RemoveContainer" containerID="661d957736179cd056778b9ec417c69d5afd27f25d079ccc3e962bd155cf48d4" Dec 27 08:05:48 crc kubenswrapper[4934]: E1227 08:05:48.410236 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-7f59f88684-gxs5t_openstack(1182d8a5-1924-45e8-b13f-18fc3b25cf9c)\"" pod="openstack/heat-cfnapi-7f59f88684-gxs5t" podUID="1182d8a5-1924-45e8-b13f-18fc3b25cf9c" Dec 27 08:05:48 crc kubenswrapper[4934]: I1227 08:05:48.412994 4934 scope.go:117] "RemoveContainer" containerID="fe3d37707fe9adabcbc2464fb446cba67915cd60ffb4fc6e35846c1b562835d2" Dec 27 08:05:48 crc kubenswrapper[4934]: E1227 08:05:48.413247 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-db89fbccf-w9xvz_openstack(7237f7ae-a705-41af-ace6-27bafbba974d)\"" pod="openstack/heat-api-db89fbccf-w9xvz" podUID="7237f7ae-a705-41af-ace6-27bafbba974d" Dec 27 08:05:48 crc kubenswrapper[4934]: I1227 08:05:48.424183 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3152d244-2397-469d-a570-d4fd09ee9a0d","Type":"ContainerStarted","Data":"59bdd4457bbe00c7bfa14dae6f351dab80c4b5fe79606c31d10b146c7fd8ffec"} Dec 27 08:05:48 crc kubenswrapper[4934]: I1227 08:05:48.431219 4934 generic.go:334] "Generic (PLEG): container finished" podID="deefee78-9e75-44e2-b72a-08aeb8b97ac4" containerID="f6b25f19435e4558d201238c1e37486a77141d8977bbfd3f0a3dfb2249350764" exitCode=2 Dec 27 08:05:48 crc kubenswrapper[4934]: I1227 08:05:48.431262 4934 generic.go:334] "Generic (PLEG): container finished" podID="deefee78-9e75-44e2-b72a-08aeb8b97ac4" containerID="ff283c40ff7fc3fffc16d3b5341ce187559ba24fdae536c0f89834ef0197cc75" exitCode=0 Dec 27 08:05:48 crc kubenswrapper[4934]: I1227 08:05:48.431447 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deefee78-9e75-44e2-b72a-08aeb8b97ac4","Type":"ContainerDied","Data":"f6b25f19435e4558d201238c1e37486a77141d8977bbfd3f0a3dfb2249350764"} Dec 27 08:05:48 crc kubenswrapper[4934]: I1227 08:05:48.431479 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deefee78-9e75-44e2-b72a-08aeb8b97ac4","Type":"ContainerDied","Data":"ff283c40ff7fc3fffc16d3b5341ce187559ba24fdae536c0f89834ef0197cc75"} Dec 27 08:05:48 crc kubenswrapper[4934]: I1227 08:05:48.480989 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.4809686810000002 podStartE2EDuration="3.480968681s" podCreationTimestamp="2025-12-27 08:05:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:05:48.471243101 +0000 UTC m=+1409.291683695" watchObservedRunningTime="2025-12-27 08:05:48.480968681 +0000 UTC m=+1409.301409275" Dec 27 08:05:49 crc kubenswrapper[4934]: I1227 08:05:49.456163 4934 scope.go:117] "RemoveContainer" containerID="fe3d37707fe9adabcbc2464fb446cba67915cd60ffb4fc6e35846c1b562835d2" Dec 27 08:05:49 crc kubenswrapper[4934]: E1227 08:05:49.456813 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-db89fbccf-w9xvz_openstack(7237f7ae-a705-41af-ace6-27bafbba974d)\"" pod="openstack/heat-api-db89fbccf-w9xvz" podUID="7237f7ae-a705-41af-ace6-27bafbba974d" Dec 27 08:05:49 crc kubenswrapper[4934]: I1227 08:05:49.457213 4934 scope.go:117] "RemoveContainer" containerID="661d957736179cd056778b9ec417c69d5afd27f25d079ccc3e962bd155cf48d4" Dec 27 08:05:49 crc kubenswrapper[4934]: E1227 08:05:49.457396 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-7f59f88684-gxs5t_openstack(1182d8a5-1924-45e8-b13f-18fc3b25cf9c)\"" pod="openstack/heat-cfnapi-7f59f88684-gxs5t" podUID="1182d8a5-1924-45e8-b13f-18fc3b25cf9c" Dec 27 08:05:49 crc kubenswrapper[4934]: I1227 08:05:49.512977 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 27 08:05:49 crc kubenswrapper[4934]: I1227 08:05:49.513224 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="86f3ebe1-9843-44e6-907a-e453aab4b9e8" containerName="glance-log" containerID="cri-o://2fd6b17ae0e178466b758988239a7b5ed70321c3d412fecab516a68a5df0d602" gracePeriod=30 Dec 27 08:05:49 crc kubenswrapper[4934]: I1227 08:05:49.513639 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="86f3ebe1-9843-44e6-907a-e453aab4b9e8" containerName="glance-httpd" containerID="cri-o://9ee83e3a8e87b6f9a43e8c2d49fec8346411b9357abba2d50e8154b7814f4de2" gracePeriod=30 Dec 27 08:05:50 crc kubenswrapper[4934]: I1227 08:05:50.472362 4934 generic.go:334] "Generic (PLEG): container finished" podID="deefee78-9e75-44e2-b72a-08aeb8b97ac4" containerID="c7aef981d05c3a0478c37d7d09c12e1d2edce81b6b397585f3f3259f03f9b58e" exitCode=0 Dec 27 08:05:50 crc kubenswrapper[4934]: I1227 08:05:50.472435 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deefee78-9e75-44e2-b72a-08aeb8b97ac4","Type":"ContainerDied","Data":"c7aef981d05c3a0478c37d7d09c12e1d2edce81b6b397585f3f3259f03f9b58e"} Dec 27 08:05:50 crc kubenswrapper[4934]: I1227 08:05:50.474896 4934 generic.go:334] "Generic (PLEG): container finished" podID="86f3ebe1-9843-44e6-907a-e453aab4b9e8" containerID="2fd6b17ae0e178466b758988239a7b5ed70321c3d412fecab516a68a5df0d602" exitCode=143 Dec 27 08:05:50 crc kubenswrapper[4934]: I1227 08:05:50.474923 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"86f3ebe1-9843-44e6-907a-e453aab4b9e8","Type":"ContainerDied","Data":"2fd6b17ae0e178466b758988239a7b5ed70321c3d412fecab516a68a5df0d602"} Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.064624 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-zkwng" Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.071049 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6kjfr" Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.079494 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xbmch" Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.147247 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tz8hp\" (UniqueName: \"kubernetes.io/projected/acbab7d1-71c2-4bf6-888b-b0e45005d374-kube-api-access-tz8hp\") pod \"acbab7d1-71c2-4bf6-888b-b0e45005d374\" (UID: \"acbab7d1-71c2-4bf6-888b-b0e45005d374\") " Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.147452 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvgfc\" (UniqueName: \"kubernetes.io/projected/8e8b0569-1879-41e6-a91c-7e1b812ff6d5-kube-api-access-gvgfc\") pod \"8e8b0569-1879-41e6-a91c-7e1b812ff6d5\" (UID: \"8e8b0569-1879-41e6-a91c-7e1b812ff6d5\") " Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.163265 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acbab7d1-71c2-4bf6-888b-b0e45005d374-kube-api-access-tz8hp" (OuterVolumeSpecName: "kube-api-access-tz8hp") pod "acbab7d1-71c2-4bf6-888b-b0e45005d374" (UID: "acbab7d1-71c2-4bf6-888b-b0e45005d374"). InnerVolumeSpecName "kube-api-access-tz8hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.166161 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e8b0569-1879-41e6-a91c-7e1b812ff6d5-kube-api-access-gvgfc" (OuterVolumeSpecName: "kube-api-access-gvgfc") pod "8e8b0569-1879-41e6-a91c-7e1b812ff6d5" (UID: "8e8b0569-1879-41e6-a91c-7e1b812ff6d5"). InnerVolumeSpecName "kube-api-access-gvgfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.249938 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wc6fr\" (UniqueName: \"kubernetes.io/projected/5cbaa946-14d7-4839-9541-772b8c5755c0-kube-api-access-wc6fr\") pod \"5cbaa946-14d7-4839-9541-772b8c5755c0\" (UID: \"5cbaa946-14d7-4839-9541-772b8c5755c0\") " Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.250524 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tz8hp\" (UniqueName: \"kubernetes.io/projected/acbab7d1-71c2-4bf6-888b-b0e45005d374-kube-api-access-tz8hp\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.250542 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvgfc\" (UniqueName: \"kubernetes.io/projected/8e8b0569-1879-41e6-a91c-7e1b812ff6d5-kube-api-access-gvgfc\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.255542 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cbaa946-14d7-4839-9541-772b8c5755c0-kube-api-access-wc6fr" (OuterVolumeSpecName: "kube-api-access-wc6fr") pod "5cbaa946-14d7-4839-9541-772b8c5755c0" (UID: "5cbaa946-14d7-4839-9541-772b8c5755c0"). InnerVolumeSpecName "kube-api-access-wc6fr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.352391 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wc6fr\" (UniqueName: \"kubernetes.io/projected/5cbaa946-14d7-4839-9541-772b8c5755c0-kube-api-access-wc6fr\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.509486 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xbmch" Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.509485 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xbmch" event={"ID":"5cbaa946-14d7-4839-9541-772b8c5755c0","Type":"ContainerDied","Data":"7918ab665d6b8f1a34e6b900208cde6cdd31943a132df593a2560a44c146a91c"} Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.509602 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7918ab665d6b8f1a34e6b900208cde6cdd31943a132df593a2560a44c146a91c" Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.511653 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6kjfr" event={"ID":"8e8b0569-1879-41e6-a91c-7e1b812ff6d5","Type":"ContainerDied","Data":"25df38bd251d365dc60d919e45b39b0af84914b7d6987dd0a7132befd03cc81b"} Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.511672 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25df38bd251d365dc60d919e45b39b0af84914b7d6987dd0a7132befd03cc81b" Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.511704 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6kjfr" Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.517489 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-zkwng" event={"ID":"acbab7d1-71c2-4bf6-888b-b0e45005d374","Type":"ContainerDied","Data":"50780ad4e074d285e8be33becca6d152a58c0f23c892747efba4785855a0da0a"} Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.517623 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50780ad4e074d285e8be33becca6d152a58c0f23c892747efba4785855a0da0a" Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.517581 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-zkwng" Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.791462 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 27 08:05:51 crc kubenswrapper[4934]: I1227 08:05:51.875604 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 27 08:05:52 crc kubenswrapper[4934]: I1227 08:05:52.526531 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c" containerName="cinder-scheduler" containerID="cri-o://cf4a8a25e30cc4b825c310997855c40903649e7ba7f340b290f00f224f9f8056" gracePeriod=30 Dec 27 08:05:52 crc kubenswrapper[4934]: I1227 08:05:52.526595 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c" containerName="probe" containerID="cri-o://a1d593c5cf64cc6fe0db885ff877caf0fcbaafeaf7c140c510a4d56580261bc5" gracePeriod=30 Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.387229 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.490914 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-db89fbccf-w9xvz"] Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.581763 4934 generic.go:334] "Generic (PLEG): container finished" podID="86f3ebe1-9843-44e6-907a-e453aab4b9e8" containerID="9ee83e3a8e87b6f9a43e8c2d49fec8346411b9357abba2d50e8154b7814f4de2" exitCode=0 Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.581832 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"86f3ebe1-9843-44e6-907a-e453aab4b9e8","Type":"ContainerDied","Data":"9ee83e3a8e87b6f9a43e8c2d49fec8346411b9357abba2d50e8154b7814f4de2"} Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.587573 4934 generic.go:334] "Generic (PLEG): container finished" podID="a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c" containerID="a1d593c5cf64cc6fe0db885ff877caf0fcbaafeaf7c140c510a4d56580261bc5" exitCode=0 Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.587640 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c","Type":"ContainerDied","Data":"a1d593c5cf64cc6fe0db885ff877caf0fcbaafeaf7c140c510a4d56580261bc5"} Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.591519 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.658505 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7f59f88684-gxs5t"] Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.705974 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.772747 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86f3ebe1-9843-44e6-907a-e453aab4b9e8-logs\") pod \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.772857 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-combined-ca-bundle\") pod \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.773351 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-config-data\") pod \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.773412 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmw7x\" (UniqueName: \"kubernetes.io/projected/86f3ebe1-9843-44e6-907a-e453aab4b9e8-kube-api-access-nmw7x\") pod \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.773503 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-internal-tls-certs\") pod \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.773635 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86f3ebe1-9843-44e6-907a-e453aab4b9e8-httpd-run\") pod \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.773666 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-scripts\") pod \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.773781 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0\") pod \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\" (UID: \"86f3ebe1-9843-44e6-907a-e453aab4b9e8\") " Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.777124 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86f3ebe1-9843-44e6-907a-e453aab4b9e8-logs" (OuterVolumeSpecName: "logs") pod "86f3ebe1-9843-44e6-907a-e453aab4b9e8" (UID: "86f3ebe1-9843-44e6-907a-e453aab4b9e8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.779359 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86f3ebe1-9843-44e6-907a-e453aab4b9e8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "86f3ebe1-9843-44e6-907a-e453aab4b9e8" (UID: "86f3ebe1-9843-44e6-907a-e453aab4b9e8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.806541 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-scripts" (OuterVolumeSpecName: "scripts") pod "86f3ebe1-9843-44e6-907a-e453aab4b9e8" (UID: "86f3ebe1-9843-44e6-907a-e453aab4b9e8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.837235 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86f3ebe1-9843-44e6-907a-e453aab4b9e8-kube-api-access-nmw7x" (OuterVolumeSpecName: "kube-api-access-nmw7x") pod "86f3ebe1-9843-44e6-907a-e453aab4b9e8" (UID: "86f3ebe1-9843-44e6-907a-e453aab4b9e8"). InnerVolumeSpecName "kube-api-access-nmw7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.849411 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0" (OuterVolumeSpecName: "glance") pod "86f3ebe1-9843-44e6-907a-e453aab4b9e8" (UID: "86f3ebe1-9843-44e6-907a-e453aab4b9e8"). InnerVolumeSpecName "pvc-44a42b08-9517-405f-9959-95693093d9c0". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.857202 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "86f3ebe1-9843-44e6-907a-e453aab4b9e8" (UID: "86f3ebe1-9843-44e6-907a-e453aab4b9e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.880479 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-config-data" (OuterVolumeSpecName: "config-data") pod "86f3ebe1-9843-44e6-907a-e453aab4b9e8" (UID: "86f3ebe1-9843-44e6-907a-e453aab4b9e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.882334 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/86f3ebe1-9843-44e6-907a-e453aab4b9e8-logs\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.882354 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.882369 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.882380 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmw7x\" (UniqueName: \"kubernetes.io/projected/86f3ebe1-9843-44e6-907a-e453aab4b9e8-kube-api-access-nmw7x\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.882391 4934 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/86f3ebe1-9843-44e6-907a-e453aab4b9e8-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.882399 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.882429 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-44a42b08-9517-405f-9959-95693093d9c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0\") on node \"crc\" " Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.882508 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-db89fbccf-w9xvz" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.923852 4934 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.924199 4934 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-44a42b08-9517-405f-9959-95693093d9c0" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0") on node "crc" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.948402 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "86f3ebe1-9843-44e6-907a-e453aab4b9e8" (UID: "86f3ebe1-9843-44e6-907a-e453aab4b9e8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.983580 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7237f7ae-a705-41af-ace6-27bafbba974d-config-data-custom\") pod \"7237f7ae-a705-41af-ace6-27bafbba974d\" (UID: \"7237f7ae-a705-41af-ace6-27bafbba974d\") " Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.983763 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8sxd6\" (UniqueName: \"kubernetes.io/projected/7237f7ae-a705-41af-ace6-27bafbba974d-kube-api-access-8sxd6\") pod \"7237f7ae-a705-41af-ace6-27bafbba974d\" (UID: \"7237f7ae-a705-41af-ace6-27bafbba974d\") " Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.983808 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7237f7ae-a705-41af-ace6-27bafbba974d-combined-ca-bundle\") pod \"7237f7ae-a705-41af-ace6-27bafbba974d\" (UID: \"7237f7ae-a705-41af-ace6-27bafbba974d\") " Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.983827 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7237f7ae-a705-41af-ace6-27bafbba974d-config-data\") pod \"7237f7ae-a705-41af-ace6-27bafbba974d\" (UID: \"7237f7ae-a705-41af-ace6-27bafbba974d\") " Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.984342 4934 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/86f3ebe1-9843-44e6-907a-e453aab4b9e8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.984359 4934 reconciler_common.go:293] "Volume detached for volume \"pvc-44a42b08-9517-405f-9959-95693093d9c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.994873 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7237f7ae-a705-41af-ace6-27bafbba974d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7237f7ae-a705-41af-ace6-27bafbba974d" (UID: "7237f7ae-a705-41af-ace6-27bafbba974d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:53 crc kubenswrapper[4934]: I1227 08:05:53.994960 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7237f7ae-a705-41af-ace6-27bafbba974d-kube-api-access-8sxd6" (OuterVolumeSpecName: "kube-api-access-8sxd6") pod "7237f7ae-a705-41af-ace6-27bafbba974d" (UID: "7237f7ae-a705-41af-ace6-27bafbba974d"). InnerVolumeSpecName "kube-api-access-8sxd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.047207 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7237f7ae-a705-41af-ace6-27bafbba974d-config-data" (OuterVolumeSpecName: "config-data") pod "7237f7ae-a705-41af-ace6-27bafbba974d" (UID: "7237f7ae-a705-41af-ace6-27bafbba974d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.052699 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7237f7ae-a705-41af-ace6-27bafbba974d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7237f7ae-a705-41af-ace6-27bafbba974d" (UID: "7237f7ae-a705-41af-ace6-27bafbba974d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.090075 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7237f7ae-a705-41af-ace6-27bafbba974d-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.090120 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8sxd6\" (UniqueName: \"kubernetes.io/projected/7237f7ae-a705-41af-ace6-27bafbba974d-kube-api-access-8sxd6\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.090129 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7237f7ae-a705-41af-ace6-27bafbba974d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.090139 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7237f7ae-a705-41af-ace6-27bafbba974d-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.260008 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7f59f88684-gxs5t" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.397153 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-config-data-custom\") pod \"1182d8a5-1924-45e8-b13f-18fc3b25cf9c\" (UID: \"1182d8a5-1924-45e8-b13f-18fc3b25cf9c\") " Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.397276 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-config-data\") pod \"1182d8a5-1924-45e8-b13f-18fc3b25cf9c\" (UID: \"1182d8a5-1924-45e8-b13f-18fc3b25cf9c\") " Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.397437 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-combined-ca-bundle\") pod \"1182d8a5-1924-45e8-b13f-18fc3b25cf9c\" (UID: \"1182d8a5-1924-45e8-b13f-18fc3b25cf9c\") " Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.397498 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jn7g\" (UniqueName: \"kubernetes.io/projected/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-kube-api-access-5jn7g\") pod \"1182d8a5-1924-45e8-b13f-18fc3b25cf9c\" (UID: \"1182d8a5-1924-45e8-b13f-18fc3b25cf9c\") " Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.404235 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-kube-api-access-5jn7g" (OuterVolumeSpecName: "kube-api-access-5jn7g") pod "1182d8a5-1924-45e8-b13f-18fc3b25cf9c" (UID: "1182d8a5-1924-45e8-b13f-18fc3b25cf9c"). InnerVolumeSpecName "kube-api-access-5jn7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.414248 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1182d8a5-1924-45e8-b13f-18fc3b25cf9c" (UID: "1182d8a5-1924-45e8-b13f-18fc3b25cf9c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.425979 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1182d8a5-1924-45e8-b13f-18fc3b25cf9c" (UID: "1182d8a5-1924-45e8-b13f-18fc3b25cf9c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.466002 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-config-data" (OuterVolumeSpecName: "config-data") pod "1182d8a5-1924-45e8-b13f-18fc3b25cf9c" (UID: "1182d8a5-1924-45e8-b13f-18fc3b25cf9c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.499496 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.499530 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.499549 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.499558 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jn7g\" (UniqueName: \"kubernetes.io/projected/1182d8a5-1924-45e8-b13f-18fc3b25cf9c-kube-api-access-5jn7g\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.601632 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-db89fbccf-w9xvz" event={"ID":"7237f7ae-a705-41af-ace6-27bafbba974d","Type":"ContainerDied","Data":"b7028b9d51c28b6cc058a2a2997853affa85d43c9ef8e0adc5df2b05a2058071"} Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.601691 4934 scope.go:117] "RemoveContainer" containerID="fe3d37707fe9adabcbc2464fb446cba67915cd60ffb4fc6e35846c1b562835d2" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.601722 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-db89fbccf-w9xvz" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.605541 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"86f3ebe1-9843-44e6-907a-e453aab4b9e8","Type":"ContainerDied","Data":"5dfe586abfcd1ecfe60596eafc9089b4922ece508d866d8b6e575effe0cbbe18"} Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.605648 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.608183 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"f5eabd4f-a103-4ec0-9c14-7deb0e36fe61","Type":"ContainerStarted","Data":"35321ccbba0f7035a140f67b6275e9df8b9edbb6e30f6b4cf7b6c0ffcd3f95e3"} Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.614198 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-7f59f88684-gxs5t" event={"ID":"1182d8a5-1924-45e8-b13f-18fc3b25cf9c","Type":"ContainerDied","Data":"56dc1ca7f2bc570dc0724c3da2a278079753d432d38acf20a66b547d90937515"} Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.614269 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-7f59f88684-gxs5t" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.630521 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.7999151960000002 podStartE2EDuration="41.630503479s" podCreationTimestamp="2025-12-27 08:05:13 +0000 UTC" firstStartedPulling="2025-12-27 08:05:14.225343143 +0000 UTC m=+1375.045783737" lastFinishedPulling="2025-12-27 08:05:54.055931426 +0000 UTC m=+1414.876372020" observedRunningTime="2025-12-27 08:05:54.629952006 +0000 UTC m=+1415.450392610" watchObservedRunningTime="2025-12-27 08:05:54.630503479 +0000 UTC m=+1415.450944073" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.657662 4934 scope.go:117] "RemoveContainer" containerID="9ee83e3a8e87b6f9a43e8c2d49fec8346411b9357abba2d50e8154b7814f4de2" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.700249 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.714363 4934 scope.go:117] "RemoveContainer" containerID="2fd6b17ae0e178466b758988239a7b5ed70321c3d412fecab516a68a5df0d602" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.722623 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.734768 4934 scope.go:117] "RemoveContainer" containerID="661d957736179cd056778b9ec417c69d5afd27f25d079ccc3e962bd155cf48d4" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.755645 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-db89fbccf-w9xvz"] Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.767493 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-db89fbccf-w9xvz"] Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.778756 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 27 08:05:54 crc kubenswrapper[4934]: E1227 08:05:54.779432 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7237f7ae-a705-41af-ace6-27bafbba974d" containerName="heat-api" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.779461 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="7237f7ae-a705-41af-ace6-27bafbba974d" containerName="heat-api" Dec 27 08:05:54 crc kubenswrapper[4934]: E1227 08:05:54.779477 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acbab7d1-71c2-4bf6-888b-b0e45005d374" containerName="mariadb-database-create" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.779485 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="acbab7d1-71c2-4bf6-888b-b0e45005d374" containerName="mariadb-database-create" Dec 27 08:05:54 crc kubenswrapper[4934]: E1227 08:05:54.779495 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e8b0569-1879-41e6-a91c-7e1b812ff6d5" containerName="mariadb-database-create" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.779502 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e8b0569-1879-41e6-a91c-7e1b812ff6d5" containerName="mariadb-database-create" Dec 27 08:05:54 crc kubenswrapper[4934]: E1227 08:05:54.779531 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cbaa946-14d7-4839-9541-772b8c5755c0" containerName="mariadb-database-create" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.779539 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cbaa946-14d7-4839-9541-772b8c5755c0" containerName="mariadb-database-create" Dec 27 08:05:54 crc kubenswrapper[4934]: E1227 08:05:54.779549 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86f3ebe1-9843-44e6-907a-e453aab4b9e8" containerName="glance-log" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.779555 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="86f3ebe1-9843-44e6-907a-e453aab4b9e8" containerName="glance-log" Dec 27 08:05:54 crc kubenswrapper[4934]: E1227 08:05:54.779586 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1182d8a5-1924-45e8-b13f-18fc3b25cf9c" containerName="heat-cfnapi" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.779596 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1182d8a5-1924-45e8-b13f-18fc3b25cf9c" containerName="heat-cfnapi" Dec 27 08:05:54 crc kubenswrapper[4934]: E1227 08:05:54.779606 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86f3ebe1-9843-44e6-907a-e453aab4b9e8" containerName="glance-httpd" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.779613 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="86f3ebe1-9843-44e6-907a-e453aab4b9e8" containerName="glance-httpd" Dec 27 08:05:54 crc kubenswrapper[4934]: E1227 08:05:54.779631 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1182d8a5-1924-45e8-b13f-18fc3b25cf9c" containerName="heat-cfnapi" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.779639 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1182d8a5-1924-45e8-b13f-18fc3b25cf9c" containerName="heat-cfnapi" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.779897 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="86f3ebe1-9843-44e6-907a-e453aab4b9e8" containerName="glance-httpd" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.779914 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cbaa946-14d7-4839-9541-772b8c5755c0" containerName="mariadb-database-create" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.779927 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="7237f7ae-a705-41af-ace6-27bafbba974d" containerName="heat-api" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.779938 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1182d8a5-1924-45e8-b13f-18fc3b25cf9c" containerName="heat-cfnapi" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.779952 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="86f3ebe1-9843-44e6-907a-e453aab4b9e8" containerName="glance-log" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.779967 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="acbab7d1-71c2-4bf6-888b-b0e45005d374" containerName="mariadb-database-create" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.780013 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e8b0569-1879-41e6-a91c-7e1b812ff6d5" containerName="mariadb-database-create" Dec 27 08:05:54 crc kubenswrapper[4934]: E1227 08:05:54.780311 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7237f7ae-a705-41af-ace6-27bafbba974d" containerName="heat-api" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.780324 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="7237f7ae-a705-41af-ace6-27bafbba974d" containerName="heat-api" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.780585 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="7237f7ae-a705-41af-ace6-27bafbba974d" containerName="heat-api" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.780606 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1182d8a5-1924-45e8-b13f-18fc3b25cf9c" containerName="heat-cfnapi" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.781680 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.784034 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.784225 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.791192 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-7f59f88684-gxs5t"] Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.802057 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-7f59f88684-gxs5t"] Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.815518 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.910881 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3966937c-e242-445a-ac07-f70a6d6a689e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.911044 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3966937c-e242-445a-ac07-f70a6d6a689e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.911165 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3966937c-e242-445a-ac07-f70a6d6a689e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.911244 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3966937c-e242-445a-ac07-f70a6d6a689e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.911281 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3966937c-e242-445a-ac07-f70a6d6a689e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.911303 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3966937c-e242-445a-ac07-f70a6d6a689e-logs\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.911384 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57pvv\" (UniqueName: \"kubernetes.io/projected/3966937c-e242-445a-ac07-f70a6d6a689e-kube-api-access-57pvv\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.911510 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-44a42b08-9517-405f-9959-95693093d9c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.990708 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-5cdf-account-create-hdhg5"] Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.992159 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5cdf-account-create-hdhg5" Dec 27 08:05:54 crc kubenswrapper[4934]: I1227 08:05:54.993913 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.001213 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5cdf-account-create-hdhg5"] Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.013441 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-44a42b08-9517-405f-9959-95693093d9c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.014135 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3966937c-e242-445a-ac07-f70a6d6a689e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.014378 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3966937c-e242-445a-ac07-f70a6d6a689e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.014898 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3966937c-e242-445a-ac07-f70a6d6a689e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.015061 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3966937c-e242-445a-ac07-f70a6d6a689e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.015305 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3966937c-e242-445a-ac07-f70a6d6a689e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.015432 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3966937c-e242-445a-ac07-f70a6d6a689e-logs\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.015599 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57pvv\" (UniqueName: \"kubernetes.io/projected/3966937c-e242-445a-ac07-f70a6d6a689e-kube-api-access-57pvv\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.016450 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3966937c-e242-445a-ac07-f70a6d6a689e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.019004 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3966937c-e242-445a-ac07-f70a6d6a689e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.021885 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.021922 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-44a42b08-9517-405f-9959-95693093d9c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/05b75074d2eec1bd1175d6431a19a8ca1381b55f1d6f90d62e5d62974d613123/globalmount\"" pod="openstack/glance-default-internal-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.022126 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3966937c-e242-445a-ac07-f70a6d6a689e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.022705 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3966937c-e242-445a-ac07-f70a6d6a689e-logs\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.024437 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3966937c-e242-445a-ac07-f70a6d6a689e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.025134 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3966937c-e242-445a-ac07-f70a6d6a689e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.034266 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57pvv\" (UniqueName: \"kubernetes.io/projected/3966937c-e242-445a-ac07-f70a6d6a689e-kube-api-access-57pvv\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.075655 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-44a42b08-9517-405f-9959-95693093d9c0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-44a42b08-9517-405f-9959-95693093d9c0\") pod \"glance-default-internal-api-0\" (UID: \"3966937c-e242-445a-ac07-f70a6d6a689e\") " pod="openstack/glance-default-internal-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.102680 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.117994 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9n7n\" (UniqueName: \"kubernetes.io/projected/0c9fcfdd-943b-42b6-9741-74feaf6d90f7-kube-api-access-t9n7n\") pod \"nova-api-5cdf-account-create-hdhg5\" (UID: \"0c9fcfdd-943b-42b6-9741-74feaf6d90f7\") " pod="openstack/nova-api-5cdf-account-create-hdhg5" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.198803 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-5b72-account-create-xsfb7"] Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.200673 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5b72-account-create-xsfb7" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.202899 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.219856 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9n7n\" (UniqueName: \"kubernetes.io/projected/0c9fcfdd-943b-42b6-9741-74feaf6d90f7-kube-api-access-t9n7n\") pod \"nova-api-5cdf-account-create-hdhg5\" (UID: \"0c9fcfdd-943b-42b6-9741-74feaf6d90f7\") " pod="openstack/nova-api-5cdf-account-create-hdhg5" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.252641 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9n7n\" (UniqueName: \"kubernetes.io/projected/0c9fcfdd-943b-42b6-9741-74feaf6d90f7-kube-api-access-t9n7n\") pod \"nova-api-5cdf-account-create-hdhg5\" (UID: \"0c9fcfdd-943b-42b6-9741-74feaf6d90f7\") " pod="openstack/nova-api-5cdf-account-create-hdhg5" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.288986 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-5b72-account-create-xsfb7"] Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.311377 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5cdf-account-create-hdhg5" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.322180 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqk94\" (UniqueName: \"kubernetes.io/projected/42a96968-f7ee-40bf-9d60-b662e888db0d-kube-api-access-jqk94\") pod \"nova-cell0-5b72-account-create-xsfb7\" (UID: \"42a96968-f7ee-40bf-9d60-b662e888db0d\") " pod="openstack/nova-cell0-5b72-account-create-xsfb7" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.395318 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-084f-account-create-9l2gc"] Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.396871 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-084f-account-create-9l2gc" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.400541 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.415833 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-084f-account-create-9l2gc"] Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.432261 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqk94\" (UniqueName: \"kubernetes.io/projected/42a96968-f7ee-40bf-9d60-b662e888db0d-kube-api-access-jqk94\") pod \"nova-cell0-5b72-account-create-xsfb7\" (UID: \"42a96968-f7ee-40bf-9d60-b662e888db0d\") " pod="openstack/nova-cell0-5b72-account-create-xsfb7" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.450958 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqk94\" (UniqueName: \"kubernetes.io/projected/42a96968-f7ee-40bf-9d60-b662e888db0d-kube-api-access-jqk94\") pod \"nova-cell0-5b72-account-create-xsfb7\" (UID: \"42a96968-f7ee-40bf-9d60-b662e888db0d\") " pod="openstack/nova-cell0-5b72-account-create-xsfb7" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.488742 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1182d8a5-1924-45e8-b13f-18fc3b25cf9c" path="/var/lib/kubelet/pods/1182d8a5-1924-45e8-b13f-18fc3b25cf9c/volumes" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.489487 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7237f7ae-a705-41af-ace6-27bafbba974d" path="/var/lib/kubelet/pods/7237f7ae-a705-41af-ace6-27bafbba974d/volumes" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.490256 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86f3ebe1-9843-44e6-907a-e453aab4b9e8" path="/var/lib/kubelet/pods/86f3ebe1-9843-44e6-907a-e453aab4b9e8/volumes" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.533835 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58bhj\" (UniqueName: \"kubernetes.io/projected/74a539ac-109e-4442-ac71-209a2ed6a27d-kube-api-access-58bhj\") pod \"nova-cell1-084f-account-create-9l2gc\" (UID: \"74a539ac-109e-4442-ac71-209a2ed6a27d\") " pod="openstack/nova-cell1-084f-account-create-9l2gc" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.557062 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.634796 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-config-data\") pod \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.635291 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whj85\" (UniqueName: \"kubernetes.io/projected/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-kube-api-access-whj85\") pod \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.635361 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-scripts\") pod \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.635412 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-etc-machine-id\") pod \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.635514 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-combined-ca-bundle\") pod \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.635569 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-config-data-custom\") pod \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\" (UID: \"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c\") " Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.636063 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58bhj\" (UniqueName: \"kubernetes.io/projected/74a539ac-109e-4442-ac71-209a2ed6a27d-kube-api-access-58bhj\") pod \"nova-cell1-084f-account-create-9l2gc\" (UID: \"74a539ac-109e-4442-ac71-209a2ed6a27d\") " pod="openstack/nova-cell1-084f-account-create-9l2gc" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.639603 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-kube-api-access-whj85" (OuterVolumeSpecName: "kube-api-access-whj85") pod "a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c" (UID: "a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c"). InnerVolumeSpecName "kube-api-access-whj85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.642781 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-scripts" (OuterVolumeSpecName: "scripts") pod "a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c" (UID: "a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.642878 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c" (UID: "a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.648447 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5b72-account-create-xsfb7" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.648673 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c" (UID: "a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.664710 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58bhj\" (UniqueName: \"kubernetes.io/projected/74a539ac-109e-4442-ac71-209a2ed6a27d-kube-api-access-58bhj\") pod \"nova-cell1-084f-account-create-9l2gc\" (UID: \"74a539ac-109e-4442-ac71-209a2ed6a27d\") " pod="openstack/nova-cell1-084f-account-create-9l2gc" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.671634 4934 generic.go:334] "Generic (PLEG): container finished" podID="a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c" containerID="cf4a8a25e30cc4b825c310997855c40903649e7ba7f340b290f00f224f9f8056" exitCode=0 Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.671676 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c","Type":"ContainerDied","Data":"cf4a8a25e30cc4b825c310997855c40903649e7ba7f340b290f00f224f9f8056"} Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.671703 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c","Type":"ContainerDied","Data":"d8722f13d72ecfef9fc81ab8b5bdbd203a46476879bb3584058882ea7edae274"} Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.671720 4934 scope.go:117] "RemoveContainer" containerID="a1d593c5cf64cc6fe0db885ff877caf0fcbaafeaf7c140c510a4d56580261bc5" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.671858 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.708996 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c" (UID: "a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.710942 4934 scope.go:117] "RemoveContainer" containerID="cf4a8a25e30cc4b825c310997855c40903649e7ba7f340b290f00f224f9f8056" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.732051 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-084f-account-create-9l2gc" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.737864 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.737887 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.737896 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whj85\" (UniqueName: \"kubernetes.io/projected/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-kube-api-access-whj85\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.737907 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.737915 4934 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.746308 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.746345 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.765809 4934 scope.go:117] "RemoveContainer" containerID="a1d593c5cf64cc6fe0db885ff877caf0fcbaafeaf7c140c510a4d56580261bc5" Dec 27 08:05:55 crc kubenswrapper[4934]: E1227 08:05:55.769184 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1d593c5cf64cc6fe0db885ff877caf0fcbaafeaf7c140c510a4d56580261bc5\": container with ID starting with a1d593c5cf64cc6fe0db885ff877caf0fcbaafeaf7c140c510a4d56580261bc5 not found: ID does not exist" containerID="a1d593c5cf64cc6fe0db885ff877caf0fcbaafeaf7c140c510a4d56580261bc5" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.769228 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1d593c5cf64cc6fe0db885ff877caf0fcbaafeaf7c140c510a4d56580261bc5"} err="failed to get container status \"a1d593c5cf64cc6fe0db885ff877caf0fcbaafeaf7c140c510a4d56580261bc5\": rpc error: code = NotFound desc = could not find container \"a1d593c5cf64cc6fe0db885ff877caf0fcbaafeaf7c140c510a4d56580261bc5\": container with ID starting with a1d593c5cf64cc6fe0db885ff877caf0fcbaafeaf7c140c510a4d56580261bc5 not found: ID does not exist" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.769256 4934 scope.go:117] "RemoveContainer" containerID="cf4a8a25e30cc4b825c310997855c40903649e7ba7f340b290f00f224f9f8056" Dec 27 08:05:55 crc kubenswrapper[4934]: E1227 08:05:55.770111 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf4a8a25e30cc4b825c310997855c40903649e7ba7f340b290f00f224f9f8056\": container with ID starting with cf4a8a25e30cc4b825c310997855c40903649e7ba7f340b290f00f224f9f8056 not found: ID does not exist" containerID="cf4a8a25e30cc4b825c310997855c40903649e7ba7f340b290f00f224f9f8056" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.770147 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf4a8a25e30cc4b825c310997855c40903649e7ba7f340b290f00f224f9f8056"} err="failed to get container status \"cf4a8a25e30cc4b825c310997855c40903649e7ba7f340b290f00f224f9f8056\": rpc error: code = NotFound desc = could not find container \"cf4a8a25e30cc4b825c310997855c40903649e7ba7f340b290f00f224f9f8056\": container with ID starting with cf4a8a25e30cc4b825c310997855c40903649e7ba7f340b290f00f224f9f8056 not found: ID does not exist" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.808576 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-config-data" (OuterVolumeSpecName: "config-data") pod "a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c" (UID: "a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.808631 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.837068 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.840456 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.841162 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 27 08:05:55 crc kubenswrapper[4934]: I1227 08:05:55.896820 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5cdf-account-create-hdhg5"] Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.049265 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.062523 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.081570 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 27 08:05:56 crc kubenswrapper[4934]: E1227 08:05:56.081978 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c" containerName="cinder-scheduler" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.081994 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c" containerName="cinder-scheduler" Dec 27 08:05:56 crc kubenswrapper[4934]: E1227 08:05:56.082030 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c" containerName="probe" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.082038 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c" containerName="probe" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.082257 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c" containerName="cinder-scheduler" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.082282 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c" containerName="probe" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.108801 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.109842 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.111598 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.134743 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-5b72-account-create-xsfb7"] Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.250671 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46b6dbac-6489-4651-8956-8c8a9767a4aa-config-data\") pod \"cinder-scheduler-0\" (UID: \"46b6dbac-6489-4651-8956-8c8a9767a4aa\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.250783 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg9xn\" (UniqueName: \"kubernetes.io/projected/46b6dbac-6489-4651-8956-8c8a9767a4aa-kube-api-access-tg9xn\") pod \"cinder-scheduler-0\" (UID: \"46b6dbac-6489-4651-8956-8c8a9767a4aa\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.250822 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46b6dbac-6489-4651-8956-8c8a9767a4aa-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"46b6dbac-6489-4651-8956-8c8a9767a4aa\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.250850 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46b6dbac-6489-4651-8956-8c8a9767a4aa-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"46b6dbac-6489-4651-8956-8c8a9767a4aa\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.250935 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46b6dbac-6489-4651-8956-8c8a9767a4aa-scripts\") pod \"cinder-scheduler-0\" (UID: \"46b6dbac-6489-4651-8956-8c8a9767a4aa\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.250962 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46b6dbac-6489-4651-8956-8c8a9767a4aa-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"46b6dbac-6489-4651-8956-8c8a9767a4aa\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.332544 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-084f-account-create-9l2gc"] Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.353590 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46b6dbac-6489-4651-8956-8c8a9767a4aa-scripts\") pod \"cinder-scheduler-0\" (UID: \"46b6dbac-6489-4651-8956-8c8a9767a4aa\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.353686 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46b6dbac-6489-4651-8956-8c8a9767a4aa-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"46b6dbac-6489-4651-8956-8c8a9767a4aa\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.353725 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46b6dbac-6489-4651-8956-8c8a9767a4aa-config-data\") pod \"cinder-scheduler-0\" (UID: \"46b6dbac-6489-4651-8956-8c8a9767a4aa\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.353890 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg9xn\" (UniqueName: \"kubernetes.io/projected/46b6dbac-6489-4651-8956-8c8a9767a4aa-kube-api-access-tg9xn\") pod \"cinder-scheduler-0\" (UID: \"46b6dbac-6489-4651-8956-8c8a9767a4aa\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.353951 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46b6dbac-6489-4651-8956-8c8a9767a4aa-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"46b6dbac-6489-4651-8956-8c8a9767a4aa\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.354005 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46b6dbac-6489-4651-8956-8c8a9767a4aa-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"46b6dbac-6489-4651-8956-8c8a9767a4aa\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.354233 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/46b6dbac-6489-4651-8956-8c8a9767a4aa-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"46b6dbac-6489-4651-8956-8c8a9767a4aa\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.370960 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46b6dbac-6489-4651-8956-8c8a9767a4aa-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"46b6dbac-6489-4651-8956-8c8a9767a4aa\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.374315 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46b6dbac-6489-4651-8956-8c8a9767a4aa-scripts\") pod \"cinder-scheduler-0\" (UID: \"46b6dbac-6489-4651-8956-8c8a9767a4aa\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.374874 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/46b6dbac-6489-4651-8956-8c8a9767a4aa-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"46b6dbac-6489-4651-8956-8c8a9767a4aa\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.375067 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46b6dbac-6489-4651-8956-8c8a9767a4aa-config-data\") pod \"cinder-scheduler-0\" (UID: \"46b6dbac-6489-4651-8956-8c8a9767a4aa\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.379017 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg9xn\" (UniqueName: \"kubernetes.io/projected/46b6dbac-6489-4651-8956-8c8a9767a4aa-kube-api-access-tg9xn\") pod \"cinder-scheduler-0\" (UID: \"46b6dbac-6489-4651-8956-8c8a9767a4aa\") " pod="openstack/cinder-scheduler-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.489481 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.704564 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5b72-account-create-xsfb7" event={"ID":"42a96968-f7ee-40bf-9d60-b662e888db0d","Type":"ContainerStarted","Data":"5dadcba05a120b34a236db90c5a70fb53fbc446fd5c98839d70ace484e18f906"} Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.704832 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5b72-account-create-xsfb7" event={"ID":"42a96968-f7ee-40bf-9d60-b662e888db0d","Type":"ContainerStarted","Data":"a608010465f3e45a82755212a89ec47a3087c5d371b8075f222032c9ae7de5c6"} Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.712114 4934 generic.go:334] "Generic (PLEG): container finished" podID="74a539ac-109e-4442-ac71-209a2ed6a27d" containerID="851f13e7d7fd0da526e524752d71b5714f8fdde4d26e110e7f6a43c94d24d2f5" exitCode=0 Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.712222 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-084f-account-create-9l2gc" event={"ID":"74a539ac-109e-4442-ac71-209a2ed6a27d","Type":"ContainerDied","Data":"851f13e7d7fd0da526e524752d71b5714f8fdde4d26e110e7f6a43c94d24d2f5"} Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.712248 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-084f-account-create-9l2gc" event={"ID":"74a539ac-109e-4442-ac71-209a2ed6a27d","Type":"ContainerStarted","Data":"f3bf5667da1e55b6c8ccfbba1c5ab9a864ee6e76b7c2df0c8ac01d74403faf90"} Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.714402 4934 generic.go:334] "Generic (PLEG): container finished" podID="0c9fcfdd-943b-42b6-9741-74feaf6d90f7" containerID="0d4b1755b557ba1c420b5022b2131c77712aced9ec65c3d60cb4378ebcc83d68" exitCode=0 Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.714477 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5cdf-account-create-hdhg5" event={"ID":"0c9fcfdd-943b-42b6-9741-74feaf6d90f7","Type":"ContainerDied","Data":"0d4b1755b557ba1c420b5022b2131c77712aced9ec65c3d60cb4378ebcc83d68"} Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.714550 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5cdf-account-create-hdhg5" event={"ID":"0c9fcfdd-943b-42b6-9741-74feaf6d90f7","Type":"ContainerStarted","Data":"eb2713f101503d6aaf5157dd8168dc12a7716d1e3932bf1412c4c88972088bc2"} Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.717134 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3966937c-e242-445a-ac07-f70a6d6a689e","Type":"ContainerStarted","Data":"294b20ae7af97fdd417515672c0c8ef107b743755693786357eabb3638cc7791"} Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.717168 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3966937c-e242-445a-ac07-f70a6d6a689e","Type":"ContainerStarted","Data":"5708925c5284db6ac0495b752da87092fa3e9c638f9b6a65709d8165e56124fe"} Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.717367 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.717474 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 27 08:05:56 crc kubenswrapper[4934]: I1227 08:05:56.982689 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 27 08:05:57 crc kubenswrapper[4934]: I1227 08:05:57.497636 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c" path="/var/lib/kubelet/pods/a3c3edbd-c0b4-4ea6-9e3a-0350f43ced2c/volumes" Dec 27 08:05:57 crc kubenswrapper[4934]: I1227 08:05:57.733073 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"3966937c-e242-445a-ac07-f70a6d6a689e","Type":"ContainerStarted","Data":"3ea5403a27851317320887de6bc22f2011255eb978d6fe6af885eb0402778630"} Dec 27 08:05:57 crc kubenswrapper[4934]: I1227 08:05:57.737503 4934 generic.go:334] "Generic (PLEG): container finished" podID="42a96968-f7ee-40bf-9d60-b662e888db0d" containerID="5dadcba05a120b34a236db90c5a70fb53fbc446fd5c98839d70ace484e18f906" exitCode=0 Dec 27 08:05:57 crc kubenswrapper[4934]: I1227 08:05:57.737603 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5b72-account-create-xsfb7" event={"ID":"42a96968-f7ee-40bf-9d60-b662e888db0d","Type":"ContainerDied","Data":"5dadcba05a120b34a236db90c5a70fb53fbc446fd5c98839d70ace484e18f906"} Dec 27 08:05:57 crc kubenswrapper[4934]: I1227 08:05:57.739459 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"46b6dbac-6489-4651-8956-8c8a9767a4aa","Type":"ContainerStarted","Data":"d9ec6c4ef059f41bc975b4a7e92f76779e051d1ebf9286f688d3659016bad76f"} Dec 27 08:05:57 crc kubenswrapper[4934]: I1227 08:05:57.761636 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.761615127 podStartE2EDuration="3.761615127s" podCreationTimestamp="2025-12-27 08:05:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:05:57.753506467 +0000 UTC m=+1418.573947061" watchObservedRunningTime="2025-12-27 08:05:57.761615127 +0000 UTC m=+1418.582055721" Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.529906 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5cdf-account-create-hdhg5" Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.533367 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5b72-account-create-xsfb7" Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.549245 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-084f-account-create-9l2gc" Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.628794 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqk94\" (UniqueName: \"kubernetes.io/projected/42a96968-f7ee-40bf-9d60-b662e888db0d-kube-api-access-jqk94\") pod \"42a96968-f7ee-40bf-9d60-b662e888db0d\" (UID: \"42a96968-f7ee-40bf-9d60-b662e888db0d\") " Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.628842 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58bhj\" (UniqueName: \"kubernetes.io/projected/74a539ac-109e-4442-ac71-209a2ed6a27d-kube-api-access-58bhj\") pod \"74a539ac-109e-4442-ac71-209a2ed6a27d\" (UID: \"74a539ac-109e-4442-ac71-209a2ed6a27d\") " Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.629012 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9n7n\" (UniqueName: \"kubernetes.io/projected/0c9fcfdd-943b-42b6-9741-74feaf6d90f7-kube-api-access-t9n7n\") pod \"0c9fcfdd-943b-42b6-9741-74feaf6d90f7\" (UID: \"0c9fcfdd-943b-42b6-9741-74feaf6d90f7\") " Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.638362 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74a539ac-109e-4442-ac71-209a2ed6a27d-kube-api-access-58bhj" (OuterVolumeSpecName: "kube-api-access-58bhj") pod "74a539ac-109e-4442-ac71-209a2ed6a27d" (UID: "74a539ac-109e-4442-ac71-209a2ed6a27d"). InnerVolumeSpecName "kube-api-access-58bhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.639888 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42a96968-f7ee-40bf-9d60-b662e888db0d-kube-api-access-jqk94" (OuterVolumeSpecName: "kube-api-access-jqk94") pod "42a96968-f7ee-40bf-9d60-b662e888db0d" (UID: "42a96968-f7ee-40bf-9d60-b662e888db0d"). InnerVolumeSpecName "kube-api-access-jqk94". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.641197 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c9fcfdd-943b-42b6-9741-74feaf6d90f7-kube-api-access-t9n7n" (OuterVolumeSpecName: "kube-api-access-t9n7n") pod "0c9fcfdd-943b-42b6-9741-74feaf6d90f7" (UID: "0c9fcfdd-943b-42b6-9741-74feaf6d90f7"). InnerVolumeSpecName "kube-api-access-t9n7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.733518 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9n7n\" (UniqueName: \"kubernetes.io/projected/0c9fcfdd-943b-42b6-9741-74feaf6d90f7-kube-api-access-t9n7n\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.733553 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqk94\" (UniqueName: \"kubernetes.io/projected/42a96968-f7ee-40bf-9d60-b662e888db0d-kube-api-access-jqk94\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.733563 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58bhj\" (UniqueName: \"kubernetes.io/projected/74a539ac-109e-4442-ac71-209a2ed6a27d-kube-api-access-58bhj\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.767347 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5b72-account-create-xsfb7" Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.768181 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5b72-account-create-xsfb7" event={"ID":"42a96968-f7ee-40bf-9d60-b662e888db0d","Type":"ContainerDied","Data":"a608010465f3e45a82755212a89ec47a3087c5d371b8075f222032c9ae7de5c6"} Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.768238 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a608010465f3e45a82755212a89ec47a3087c5d371b8075f222032c9ae7de5c6" Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.777702 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"46b6dbac-6489-4651-8956-8c8a9767a4aa","Type":"ContainerStarted","Data":"d56b14937b8f4e2616b25041792887ddd2d75019777c58f0718b35781615cd6b"} Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.794394 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-084f-account-create-9l2gc" event={"ID":"74a539ac-109e-4442-ac71-209a2ed6a27d","Type":"ContainerDied","Data":"f3bf5667da1e55b6c8ccfbba1c5ab9a864ee6e76b7c2df0c8ac01d74403faf90"} Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.794430 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3bf5667da1e55b6c8ccfbba1c5ab9a864ee6e76b7c2df0c8ac01d74403faf90" Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.794484 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-084f-account-create-9l2gc" Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.800605 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5cdf-account-create-hdhg5" Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.802162 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5cdf-account-create-hdhg5" event={"ID":"0c9fcfdd-943b-42b6-9741-74feaf6d90f7","Type":"ContainerDied","Data":"eb2713f101503d6aaf5157dd8168dc12a7716d1e3932bf1412c4c88972088bc2"} Dec 27 08:05:58 crc kubenswrapper[4934]: I1227 08:05:58.802206 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb2713f101503d6aaf5157dd8168dc12a7716d1e3932bf1412c4c88972088bc2" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.202451 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.348766 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-config-data\") pod \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.348857 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/34c0fb07-e3e3-45f0-b785-5c981a771ceb-etc-machine-id\") pod \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.348931 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-config-data-custom\") pod \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.348956 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdhlk\" (UniqueName: \"kubernetes.io/projected/34c0fb07-e3e3-45f0-b785-5c981a771ceb-kube-api-access-hdhlk\") pod \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.348991 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34c0fb07-e3e3-45f0-b785-5c981a771ceb-logs\") pod \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.349021 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-scripts\") pod \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.349075 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-combined-ca-bundle\") pod \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\" (UID: \"34c0fb07-e3e3-45f0-b785-5c981a771ceb\") " Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.350662 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34c0fb07-e3e3-45f0-b785-5c981a771ceb-logs" (OuterVolumeSpecName: "logs") pod "34c0fb07-e3e3-45f0-b785-5c981a771ceb" (UID: "34c0fb07-e3e3-45f0-b785-5c981a771ceb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.357760 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34c0fb07-e3e3-45f0-b785-5c981a771ceb-kube-api-access-hdhlk" (OuterVolumeSpecName: "kube-api-access-hdhlk") pod "34c0fb07-e3e3-45f0-b785-5c981a771ceb" (UID: "34c0fb07-e3e3-45f0-b785-5c981a771ceb"). InnerVolumeSpecName "kube-api-access-hdhlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.358012 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/34c0fb07-e3e3-45f0-b785-5c981a771ceb-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "34c0fb07-e3e3-45f0-b785-5c981a771ceb" (UID: "34c0fb07-e3e3-45f0-b785-5c981a771ceb"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.370268 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "34c0fb07-e3e3-45f0-b785-5c981a771ceb" (UID: "34c0fb07-e3e3-45f0-b785-5c981a771ceb"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.373494 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-scripts" (OuterVolumeSpecName: "scripts") pod "34c0fb07-e3e3-45f0-b785-5c981a771ceb" (UID: "34c0fb07-e3e3-45f0-b785-5c981a771ceb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.454704 4934 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/34c0fb07-e3e3-45f0-b785-5c981a771ceb-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.454743 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.454753 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdhlk\" (UniqueName: \"kubernetes.io/projected/34c0fb07-e3e3-45f0-b785-5c981a771ceb-kube-api-access-hdhlk\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.454765 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/34c0fb07-e3e3-45f0-b785-5c981a771ceb-logs\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.454773 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.528487 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "34c0fb07-e3e3-45f0-b785-5c981a771ceb" (UID: "34c0fb07-e3e3-45f0-b785-5c981a771ceb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.564004 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-config-data" (OuterVolumeSpecName: "config-data") pod "34c0fb07-e3e3-45f0-b785-5c981a771ceb" (UID: "34c0fb07-e3e3-45f0-b785-5c981a771ceb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.587261 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.587301 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34c0fb07-e3e3-45f0-b785-5c981a771ceb-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.806749 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.806868 4934 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.820076 4934 generic.go:334] "Generic (PLEG): container finished" podID="34c0fb07-e3e3-45f0-b785-5c981a771ceb" containerID="958bfd98687130b5348a39978c055d6505574984ccbf8cb3b6e95825cf5dc1c7" exitCode=137 Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.820194 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"34c0fb07-e3e3-45f0-b785-5c981a771ceb","Type":"ContainerDied","Data":"958bfd98687130b5348a39978c055d6505574984ccbf8cb3b6e95825cf5dc1c7"} Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.820225 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"34c0fb07-e3e3-45f0-b785-5c981a771ceb","Type":"ContainerDied","Data":"d8970be25c26fa6cdfb0d68853daf1b63993f7f12ebadb265a2136b3771992f4"} Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.820243 4934 scope.go:117] "RemoveContainer" containerID="958bfd98687130b5348a39978c055d6505574984ccbf8cb3b6e95825cf5dc1c7" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.820436 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.823688 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"46b6dbac-6489-4651-8956-8c8a9767a4aa","Type":"ContainerStarted","Data":"9a755b3b9d9e9af687c8c817f89ff94f9621b3d56a5e91a6cf80a545b91e9b58"} Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.853852 4934 scope.go:117] "RemoveContainer" containerID="6ba9e0de8f593334de9c3a3f840f5b826c635cfefee0191b276a6548829c4f4b" Dec 27 08:05:59 crc kubenswrapper[4934]: E1227 08:05:59.857937 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34c0fb07_e3e3_45f0_b785_5c981a771ceb.slice/crio-d8970be25c26fa6cdfb0d68853daf1b63993f7f12ebadb265a2136b3771992f4\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34c0fb07_e3e3_45f0_b785_5c981a771ceb.slice\": RecentStats: unable to find data in memory cache]" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.858303 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.8582848050000003 podStartE2EDuration="3.858284805s" podCreationTimestamp="2025-12-27 08:05:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:05:59.854642475 +0000 UTC m=+1420.675083069" watchObservedRunningTime="2025-12-27 08:05:59.858284805 +0000 UTC m=+1420.678725399" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.886157 4934 scope.go:117] "RemoveContainer" containerID="958bfd98687130b5348a39978c055d6505574984ccbf8cb3b6e95825cf5dc1c7" Dec 27 08:05:59 crc kubenswrapper[4934]: E1227 08:05:59.886875 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"958bfd98687130b5348a39978c055d6505574984ccbf8cb3b6e95825cf5dc1c7\": container with ID starting with 958bfd98687130b5348a39978c055d6505574984ccbf8cb3b6e95825cf5dc1c7 not found: ID does not exist" containerID="958bfd98687130b5348a39978c055d6505574984ccbf8cb3b6e95825cf5dc1c7" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.886934 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"958bfd98687130b5348a39978c055d6505574984ccbf8cb3b6e95825cf5dc1c7"} err="failed to get container status \"958bfd98687130b5348a39978c055d6505574984ccbf8cb3b6e95825cf5dc1c7\": rpc error: code = NotFound desc = could not find container \"958bfd98687130b5348a39978c055d6505574984ccbf8cb3b6e95825cf5dc1c7\": container with ID starting with 958bfd98687130b5348a39978c055d6505574984ccbf8cb3b6e95825cf5dc1c7 not found: ID does not exist" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.886958 4934 scope.go:117] "RemoveContainer" containerID="6ba9e0de8f593334de9c3a3f840f5b826c635cfefee0191b276a6548829c4f4b" Dec 27 08:05:59 crc kubenswrapper[4934]: E1227 08:05:59.887181 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ba9e0de8f593334de9c3a3f840f5b826c635cfefee0191b276a6548829c4f4b\": container with ID starting with 6ba9e0de8f593334de9c3a3f840f5b826c635cfefee0191b276a6548829c4f4b not found: ID does not exist" containerID="6ba9e0de8f593334de9c3a3f840f5b826c635cfefee0191b276a6548829c4f4b" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.887212 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ba9e0de8f593334de9c3a3f840f5b826c635cfefee0191b276a6548829c4f4b"} err="failed to get container status \"6ba9e0de8f593334de9c3a3f840f5b826c635cfefee0191b276a6548829c4f4b\": rpc error: code = NotFound desc = could not find container \"6ba9e0de8f593334de9c3a3f840f5b826c635cfefee0191b276a6548829c4f4b\": container with ID starting with 6ba9e0de8f593334de9c3a3f840f5b826c635cfefee0191b276a6548829c4f4b not found: ID does not exist" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.901440 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.912824 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.922795 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 27 08:05:59 crc kubenswrapper[4934]: E1227 08:05:59.923562 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42a96968-f7ee-40bf-9d60-b662e888db0d" containerName="mariadb-account-create" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.923644 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="42a96968-f7ee-40bf-9d60-b662e888db0d" containerName="mariadb-account-create" Dec 27 08:05:59 crc kubenswrapper[4934]: E1227 08:05:59.923719 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c0fb07-e3e3-45f0-b785-5c981a771ceb" containerName="cinder-api" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.923779 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c0fb07-e3e3-45f0-b785-5c981a771ceb" containerName="cinder-api" Dec 27 08:05:59 crc kubenswrapper[4934]: E1227 08:05:59.923840 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34c0fb07-e3e3-45f0-b785-5c981a771ceb" containerName="cinder-api-log" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.923889 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="34c0fb07-e3e3-45f0-b785-5c981a771ceb" containerName="cinder-api-log" Dec 27 08:05:59 crc kubenswrapper[4934]: E1227 08:05:59.923948 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c9fcfdd-943b-42b6-9741-74feaf6d90f7" containerName="mariadb-account-create" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.924004 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c9fcfdd-943b-42b6-9741-74feaf6d90f7" containerName="mariadb-account-create" Dec 27 08:05:59 crc kubenswrapper[4934]: E1227 08:05:59.924064 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a539ac-109e-4442-ac71-209a2ed6a27d" containerName="mariadb-account-create" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.924140 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a539ac-109e-4442-ac71-209a2ed6a27d" containerName="mariadb-account-create" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.924435 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="34c0fb07-e3e3-45f0-b785-5c981a771ceb" containerName="cinder-api-log" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.924497 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c9fcfdd-943b-42b6-9741-74feaf6d90f7" containerName="mariadb-account-create" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.924569 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="74a539ac-109e-4442-ac71-209a2ed6a27d" containerName="mariadb-account-create" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.924631 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="34c0fb07-e3e3-45f0-b785-5c981a771ceb" containerName="cinder-api" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.924698 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="42a96968-f7ee-40bf-9d60-b662e888db0d" containerName="mariadb-account-create" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.926871 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.930733 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.931025 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.931364 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.932449 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 27 08:05:59 crc kubenswrapper[4934]: I1227 08:05:59.936050 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.007968 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3981f015-fd8d-4429-b900-ada95f9fffa9-config-data\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.008340 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lvqf\" (UniqueName: \"kubernetes.io/projected/3981f015-fd8d-4429-b900-ada95f9fffa9-kube-api-access-8lvqf\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.008469 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3981f015-fd8d-4429-b900-ada95f9fffa9-scripts\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.008621 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3981f015-fd8d-4429-b900-ada95f9fffa9-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.008728 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3981f015-fd8d-4429-b900-ada95f9fffa9-public-tls-certs\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.008770 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3981f015-fd8d-4429-b900-ada95f9fffa9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.008794 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3981f015-fd8d-4429-b900-ada95f9fffa9-logs\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.008840 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3981f015-fd8d-4429-b900-ada95f9fffa9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.008861 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3981f015-fd8d-4429-b900-ada95f9fffa9-config-data-custom\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.111541 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lvqf\" (UniqueName: \"kubernetes.io/projected/3981f015-fd8d-4429-b900-ada95f9fffa9-kube-api-access-8lvqf\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.111600 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3981f015-fd8d-4429-b900-ada95f9fffa9-scripts\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.111644 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3981f015-fd8d-4429-b900-ada95f9fffa9-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.111686 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3981f015-fd8d-4429-b900-ada95f9fffa9-public-tls-certs\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.111705 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3981f015-fd8d-4429-b900-ada95f9fffa9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.111726 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3981f015-fd8d-4429-b900-ada95f9fffa9-logs\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.111749 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3981f015-fd8d-4429-b900-ada95f9fffa9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.111770 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3981f015-fd8d-4429-b900-ada95f9fffa9-config-data-custom\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.111844 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3981f015-fd8d-4429-b900-ada95f9fffa9-config-data\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.112525 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3981f015-fd8d-4429-b900-ada95f9fffa9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.112818 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3981f015-fd8d-4429-b900-ada95f9fffa9-logs\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.115787 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3981f015-fd8d-4429-b900-ada95f9fffa9-config-data-custom\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.118020 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3981f015-fd8d-4429-b900-ada95f9fffa9-scripts\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.118075 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3981f015-fd8d-4429-b900-ada95f9fffa9-config-data\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.118775 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3981f015-fd8d-4429-b900-ada95f9fffa9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.126929 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3981f015-fd8d-4429-b900-ada95f9fffa9-public-tls-certs\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.127054 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3981f015-fd8d-4429-b900-ada95f9fffa9-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.130573 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lvqf\" (UniqueName: \"kubernetes.io/projected/3981f015-fd8d-4429-b900-ada95f9fffa9-kube-api-access-8lvqf\") pod \"cinder-api-0\" (UID: \"3981f015-fd8d-4429-b900-ada95f9fffa9\") " pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.244153 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.451417 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cm8ph"] Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.453201 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-cm8ph" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.455016 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.455213 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-tlgks" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.455263 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.476931 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cm8ph"] Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.621937 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85e47792-15be-4b6b-8a8e-8c05c4ae4605-scripts\") pod \"nova-cell0-conductor-db-sync-cm8ph\" (UID: \"85e47792-15be-4b6b-8a8e-8c05c4ae4605\") " pod="openstack/nova-cell0-conductor-db-sync-cm8ph" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.622012 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrxd9\" (UniqueName: \"kubernetes.io/projected/85e47792-15be-4b6b-8a8e-8c05c4ae4605-kube-api-access-hrxd9\") pod \"nova-cell0-conductor-db-sync-cm8ph\" (UID: \"85e47792-15be-4b6b-8a8e-8c05c4ae4605\") " pod="openstack/nova-cell0-conductor-db-sync-cm8ph" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.622864 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85e47792-15be-4b6b-8a8e-8c05c4ae4605-config-data\") pod \"nova-cell0-conductor-db-sync-cm8ph\" (UID: \"85e47792-15be-4b6b-8a8e-8c05c4ae4605\") " pod="openstack/nova-cell0-conductor-db-sync-cm8ph" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.622949 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85e47792-15be-4b6b-8a8e-8c05c4ae4605-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-cm8ph\" (UID: \"85e47792-15be-4b6b-8a8e-8c05c4ae4605\") " pod="openstack/nova-cell0-conductor-db-sync-cm8ph" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.724763 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85e47792-15be-4b6b-8a8e-8c05c4ae4605-config-data\") pod \"nova-cell0-conductor-db-sync-cm8ph\" (UID: \"85e47792-15be-4b6b-8a8e-8c05c4ae4605\") " pod="openstack/nova-cell0-conductor-db-sync-cm8ph" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.725539 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85e47792-15be-4b6b-8a8e-8c05c4ae4605-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-cm8ph\" (UID: \"85e47792-15be-4b6b-8a8e-8c05c4ae4605\") " pod="openstack/nova-cell0-conductor-db-sync-cm8ph" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.725640 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85e47792-15be-4b6b-8a8e-8c05c4ae4605-scripts\") pod \"nova-cell0-conductor-db-sync-cm8ph\" (UID: \"85e47792-15be-4b6b-8a8e-8c05c4ae4605\") " pod="openstack/nova-cell0-conductor-db-sync-cm8ph" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.725686 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrxd9\" (UniqueName: \"kubernetes.io/projected/85e47792-15be-4b6b-8a8e-8c05c4ae4605-kube-api-access-hrxd9\") pod \"nova-cell0-conductor-db-sync-cm8ph\" (UID: \"85e47792-15be-4b6b-8a8e-8c05c4ae4605\") " pod="openstack/nova-cell0-conductor-db-sync-cm8ph" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.730693 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85e47792-15be-4b6b-8a8e-8c05c4ae4605-config-data\") pod \"nova-cell0-conductor-db-sync-cm8ph\" (UID: \"85e47792-15be-4b6b-8a8e-8c05c4ae4605\") " pod="openstack/nova-cell0-conductor-db-sync-cm8ph" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.744905 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrxd9\" (UniqueName: \"kubernetes.io/projected/85e47792-15be-4b6b-8a8e-8c05c4ae4605-kube-api-access-hrxd9\") pod \"nova-cell0-conductor-db-sync-cm8ph\" (UID: \"85e47792-15be-4b6b-8a8e-8c05c4ae4605\") " pod="openstack/nova-cell0-conductor-db-sync-cm8ph" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.744912 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85e47792-15be-4b6b-8a8e-8c05c4ae4605-scripts\") pod \"nova-cell0-conductor-db-sync-cm8ph\" (UID: \"85e47792-15be-4b6b-8a8e-8c05c4ae4605\") " pod="openstack/nova-cell0-conductor-db-sync-cm8ph" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.746581 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85e47792-15be-4b6b-8a8e-8c05c4ae4605-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-cm8ph\" (UID: \"85e47792-15be-4b6b-8a8e-8c05c4ae4605\") " pod="openstack/nova-cell0-conductor-db-sync-cm8ph" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.781481 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-cm8ph" Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.794421 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 27 08:06:00 crc kubenswrapper[4934]: W1227 08:06:00.809543 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3981f015_fd8d_4429_b900_ada95f9fffa9.slice/crio-74d319c6ca37e71b32d7c497da420dcac653f79fba4501ad34afebe1e6a355d2 WatchSource:0}: Error finding container 74d319c6ca37e71b32d7c497da420dcac653f79fba4501ad34afebe1e6a355d2: Status 404 returned error can't find the container with id 74d319c6ca37e71b32d7c497da420dcac653f79fba4501ad34afebe1e6a355d2 Dec 27 08:06:00 crc kubenswrapper[4934]: I1227 08:06:00.856993 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3981f015-fd8d-4429-b900-ada95f9fffa9","Type":"ContainerStarted","Data":"74d319c6ca37e71b32d7c497da420dcac653f79fba4501ad34afebe1e6a355d2"} Dec 27 08:06:01 crc kubenswrapper[4934]: I1227 08:06:01.298878 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cm8ph"] Dec 27 08:06:01 crc kubenswrapper[4934]: W1227 08:06:01.307067 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85e47792_15be_4b6b_8a8e_8c05c4ae4605.slice/crio-c7bc93e79f7230a8abc78e7e83b917dbb13c98c0af36e810d676fb4075b8a1d7 WatchSource:0}: Error finding container c7bc93e79f7230a8abc78e7e83b917dbb13c98c0af36e810d676fb4075b8a1d7: Status 404 returned error can't find the container with id c7bc93e79f7230a8abc78e7e83b917dbb13c98c0af36e810d676fb4075b8a1d7 Dec 27 08:06:01 crc kubenswrapper[4934]: I1227 08:06:01.492285 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34c0fb07-e3e3-45f0-b785-5c981a771ceb" path="/var/lib/kubelet/pods/34c0fb07-e3e3-45f0-b785-5c981a771ceb/volumes" Dec 27 08:06:01 crc kubenswrapper[4934]: I1227 08:06:01.493565 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 27 08:06:01 crc kubenswrapper[4934]: I1227 08:06:01.873351 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3981f015-fd8d-4429-b900-ada95f9fffa9","Type":"ContainerStarted","Data":"2d047fe139cb83a950a49969b8e2830c99ebb4831753948d50b5d58700f515a2"} Dec 27 08:06:01 crc kubenswrapper[4934]: I1227 08:06:01.876673 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-cm8ph" event={"ID":"85e47792-15be-4b6b-8a8e-8c05c4ae4605","Type":"ContainerStarted","Data":"c7bc93e79f7230a8abc78e7e83b917dbb13c98c0af36e810d676fb4075b8a1d7"} Dec 27 08:06:02 crc kubenswrapper[4934]: I1227 08:06:02.791662 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-6f98b67c6-55vgd" Dec 27 08:06:02 crc kubenswrapper[4934]: I1227 08:06:02.850503 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-6bdc8c9885-7z5pz"] Dec 27 08:06:02 crc kubenswrapper[4934]: I1227 08:06:02.850712 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-6bdc8c9885-7z5pz" podUID="2c8feac2-c2ee-484c-baed-052bd7c0b468" containerName="heat-engine" containerID="cri-o://e9bfb9c690528fcbf2b4f8d9a634e42f80fbf609c10f014e8dd8b79e094981af" gracePeriod=60 Dec 27 08:06:02 crc kubenswrapper[4934]: I1227 08:06:02.909973 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"3981f015-fd8d-4429-b900-ada95f9fffa9","Type":"ContainerStarted","Data":"330d2f0b6a10d3ae7e5377ae5d20da90bb46c49c3b83d0a3be443e8bfc7b3b7c"} Dec 27 08:06:02 crc kubenswrapper[4934]: I1227 08:06:02.910385 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 27 08:06:02 crc kubenswrapper[4934]: I1227 08:06:02.960556 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.960533811 podStartE2EDuration="3.960533811s" podCreationTimestamp="2025-12-27 08:05:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:06:02.934351546 +0000 UTC m=+1423.754792140" watchObservedRunningTime="2025-12-27 08:06:02.960533811 +0000 UTC m=+1423.780974405" Dec 27 08:06:05 crc kubenswrapper[4934]: I1227 08:06:05.104411 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 27 08:06:05 crc kubenswrapper[4934]: I1227 08:06:05.104678 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 27 08:06:05 crc kubenswrapper[4934]: I1227 08:06:05.151246 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 27 08:06:05 crc kubenswrapper[4934]: I1227 08:06:05.166950 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 27 08:06:05 crc kubenswrapper[4934]: I1227 08:06:05.946145 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 27 08:06:05 crc kubenswrapper[4934]: I1227 08:06:05.946215 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 27 08:06:06 crc kubenswrapper[4934]: I1227 08:06:06.834246 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 27 08:06:08 crc kubenswrapper[4934]: I1227 08:06:08.446021 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 27 08:06:08 crc kubenswrapper[4934]: I1227 08:06:08.446748 4934 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 27 08:06:08 crc kubenswrapper[4934]: I1227 08:06:08.453763 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 27 08:06:12 crc kubenswrapper[4934]: I1227 08:06:12.035419 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-cm8ph" event={"ID":"85e47792-15be-4b6b-8a8e-8c05c4ae4605","Type":"ContainerStarted","Data":"a25540b169ccd286d2daa333158ae981f01dc148e21953b9855b096d30040eb1"} Dec 27 08:06:12 crc kubenswrapper[4934]: I1227 08:06:12.050858 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-cm8ph" podStartSLOduration=1.745505438 podStartE2EDuration="12.050841205s" podCreationTimestamp="2025-12-27 08:06:00 +0000 UTC" firstStartedPulling="2025-12-27 08:06:01.30990049 +0000 UTC m=+1422.130341084" lastFinishedPulling="2025-12-27 08:06:11.615236247 +0000 UTC m=+1432.435676851" observedRunningTime="2025-12-27 08:06:12.048976319 +0000 UTC m=+1432.869416913" watchObservedRunningTime="2025-12-27 08:06:12.050841205 +0000 UTC m=+1432.871281799" Dec 27 08:06:12 crc kubenswrapper[4934]: E1227 08:06:12.436053 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e9bfb9c690528fcbf2b4f8d9a634e42f80fbf609c10f014e8dd8b79e094981af" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 27 08:06:12 crc kubenswrapper[4934]: E1227 08:06:12.437530 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e9bfb9c690528fcbf2b4f8d9a634e42f80fbf609c10f014e8dd8b79e094981af" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 27 08:06:12 crc kubenswrapper[4934]: E1227 08:06:12.438914 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e9bfb9c690528fcbf2b4f8d9a634e42f80fbf609c10f014e8dd8b79e094981af" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 27 08:06:12 crc kubenswrapper[4934]: E1227 08:06:12.438973 4934 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-6bdc8c9885-7z5pz" podUID="2c8feac2-c2ee-484c-baed-052bd7c0b468" containerName="heat-engine" Dec 27 08:06:12 crc kubenswrapper[4934]: I1227 08:06:12.591600 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 27 08:06:15 crc kubenswrapper[4934]: I1227 08:06:15.876005 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6bdc8c9885-7z5pz" Dec 27 08:06:16 crc kubenswrapper[4934]: I1227 08:06:16.000423 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c8feac2-c2ee-484c-baed-052bd7c0b468-config-data\") pod \"2c8feac2-c2ee-484c-baed-052bd7c0b468\" (UID: \"2c8feac2-c2ee-484c-baed-052bd7c0b468\") " Dec 27 08:06:16 crc kubenswrapper[4934]: I1227 08:06:16.000736 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctnlb\" (UniqueName: \"kubernetes.io/projected/2c8feac2-c2ee-484c-baed-052bd7c0b468-kube-api-access-ctnlb\") pod \"2c8feac2-c2ee-484c-baed-052bd7c0b468\" (UID: \"2c8feac2-c2ee-484c-baed-052bd7c0b468\") " Dec 27 08:06:16 crc kubenswrapper[4934]: I1227 08:06:16.000899 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c8feac2-c2ee-484c-baed-052bd7c0b468-combined-ca-bundle\") pod \"2c8feac2-c2ee-484c-baed-052bd7c0b468\" (UID: \"2c8feac2-c2ee-484c-baed-052bd7c0b468\") " Dec 27 08:06:16 crc kubenswrapper[4934]: I1227 08:06:16.000962 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2c8feac2-c2ee-484c-baed-052bd7c0b468-config-data-custom\") pod \"2c8feac2-c2ee-484c-baed-052bd7c0b468\" (UID: \"2c8feac2-c2ee-484c-baed-052bd7c0b468\") " Dec 27 08:06:16 crc kubenswrapper[4934]: I1227 08:06:16.006870 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c8feac2-c2ee-484c-baed-052bd7c0b468-kube-api-access-ctnlb" (OuterVolumeSpecName: "kube-api-access-ctnlb") pod "2c8feac2-c2ee-484c-baed-052bd7c0b468" (UID: "2c8feac2-c2ee-484c-baed-052bd7c0b468"). InnerVolumeSpecName "kube-api-access-ctnlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:06:16 crc kubenswrapper[4934]: I1227 08:06:16.020641 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c8feac2-c2ee-484c-baed-052bd7c0b468-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2c8feac2-c2ee-484c-baed-052bd7c0b468" (UID: "2c8feac2-c2ee-484c-baed-052bd7c0b468"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:16 crc kubenswrapper[4934]: I1227 08:06:16.032719 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c8feac2-c2ee-484c-baed-052bd7c0b468-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c8feac2-c2ee-484c-baed-052bd7c0b468" (UID: "2c8feac2-c2ee-484c-baed-052bd7c0b468"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:16 crc kubenswrapper[4934]: I1227 08:06:16.058436 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c8feac2-c2ee-484c-baed-052bd7c0b468-config-data" (OuterVolumeSpecName: "config-data") pod "2c8feac2-c2ee-484c-baed-052bd7c0b468" (UID: "2c8feac2-c2ee-484c-baed-052bd7c0b468"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:16 crc kubenswrapper[4934]: I1227 08:06:16.084391 4934 generic.go:334] "Generic (PLEG): container finished" podID="2c8feac2-c2ee-484c-baed-052bd7c0b468" containerID="e9bfb9c690528fcbf2b4f8d9a634e42f80fbf609c10f014e8dd8b79e094981af" exitCode=0 Dec 27 08:06:16 crc kubenswrapper[4934]: I1227 08:06:16.084430 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6bdc8c9885-7z5pz" event={"ID":"2c8feac2-c2ee-484c-baed-052bd7c0b468","Type":"ContainerDied","Data":"e9bfb9c690528fcbf2b4f8d9a634e42f80fbf609c10f014e8dd8b79e094981af"} Dec 27 08:06:16 crc kubenswrapper[4934]: I1227 08:06:16.084455 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6bdc8c9885-7z5pz" event={"ID":"2c8feac2-c2ee-484c-baed-052bd7c0b468","Type":"ContainerDied","Data":"3dc9a8c016fe3a231a4ad11a2d12e3bbec94ea7f70c4f3e3924310c286633dd3"} Dec 27 08:06:16 crc kubenswrapper[4934]: I1227 08:06:16.084471 4934 scope.go:117] "RemoveContainer" containerID="e9bfb9c690528fcbf2b4f8d9a634e42f80fbf609c10f014e8dd8b79e094981af" Dec 27 08:06:16 crc kubenswrapper[4934]: I1227 08:06:16.084586 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6bdc8c9885-7z5pz" Dec 27 08:06:16 crc kubenswrapper[4934]: I1227 08:06:16.103475 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c8feac2-c2ee-484c-baed-052bd7c0b468-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:16 crc kubenswrapper[4934]: I1227 08:06:16.103521 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctnlb\" (UniqueName: \"kubernetes.io/projected/2c8feac2-c2ee-484c-baed-052bd7c0b468-kube-api-access-ctnlb\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:16 crc kubenswrapper[4934]: I1227 08:06:16.103538 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c8feac2-c2ee-484c-baed-052bd7c0b468-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:16 crc kubenswrapper[4934]: I1227 08:06:16.103553 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2c8feac2-c2ee-484c-baed-052bd7c0b468-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:16 crc kubenswrapper[4934]: I1227 08:06:16.153346 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-6bdc8c9885-7z5pz"] Dec 27 08:06:16 crc kubenswrapper[4934]: I1227 08:06:16.164087 4934 scope.go:117] "RemoveContainer" containerID="e9bfb9c690528fcbf2b4f8d9a634e42f80fbf609c10f014e8dd8b79e094981af" Dec 27 08:06:16 crc kubenswrapper[4934]: E1227 08:06:16.164668 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9bfb9c690528fcbf2b4f8d9a634e42f80fbf609c10f014e8dd8b79e094981af\": container with ID starting with e9bfb9c690528fcbf2b4f8d9a634e42f80fbf609c10f014e8dd8b79e094981af not found: ID does not exist" containerID="e9bfb9c690528fcbf2b4f8d9a634e42f80fbf609c10f014e8dd8b79e094981af" Dec 27 08:06:16 crc kubenswrapper[4934]: I1227 08:06:16.164711 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9bfb9c690528fcbf2b4f8d9a634e42f80fbf609c10f014e8dd8b79e094981af"} err="failed to get container status \"e9bfb9c690528fcbf2b4f8d9a634e42f80fbf609c10f014e8dd8b79e094981af\": rpc error: code = NotFound desc = could not find container \"e9bfb9c690528fcbf2b4f8d9a634e42f80fbf609c10f014e8dd8b79e094981af\": container with ID starting with e9bfb9c690528fcbf2b4f8d9a634e42f80fbf609c10f014e8dd8b79e094981af not found: ID does not exist" Dec 27 08:06:16 crc kubenswrapper[4934]: I1227 08:06:16.167535 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-6bdc8c9885-7z5pz"] Dec 27 08:06:17 crc kubenswrapper[4934]: I1227 08:06:17.483596 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c8feac2-c2ee-484c-baed-052bd7c0b468" path="/var/lib/kubelet/pods/2c8feac2-c2ee-484c-baed-052bd7c0b468/volumes" Dec 27 08:06:17 crc kubenswrapper[4934]: I1227 08:06:17.816434 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:06:17 crc kubenswrapper[4934]: I1227 08:06:17.947390 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-config-data\") pod \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " Dec 27 08:06:17 crc kubenswrapper[4934]: I1227 08:06:17.947510 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-sg-core-conf-yaml\") pod \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " Dec 27 08:06:17 crc kubenswrapper[4934]: I1227 08:06:17.947584 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-combined-ca-bundle\") pod \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " Dec 27 08:06:17 crc kubenswrapper[4934]: I1227 08:06:17.947696 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55fc2\" (UniqueName: \"kubernetes.io/projected/deefee78-9e75-44e2-b72a-08aeb8b97ac4-kube-api-access-55fc2\") pod \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " Dec 27 08:06:17 crc kubenswrapper[4934]: I1227 08:06:17.947748 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deefee78-9e75-44e2-b72a-08aeb8b97ac4-run-httpd\") pod \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " Dec 27 08:06:17 crc kubenswrapper[4934]: I1227 08:06:17.947826 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-scripts\") pod \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " Dec 27 08:06:17 crc kubenswrapper[4934]: I1227 08:06:17.947969 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deefee78-9e75-44e2-b72a-08aeb8b97ac4-log-httpd\") pod \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\" (UID: \"deefee78-9e75-44e2-b72a-08aeb8b97ac4\") " Dec 27 08:06:17 crc kubenswrapper[4934]: I1227 08:06:17.948827 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/deefee78-9e75-44e2-b72a-08aeb8b97ac4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "deefee78-9e75-44e2-b72a-08aeb8b97ac4" (UID: "deefee78-9e75-44e2-b72a-08aeb8b97ac4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:06:17 crc kubenswrapper[4934]: I1227 08:06:17.949448 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/deefee78-9e75-44e2-b72a-08aeb8b97ac4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "deefee78-9e75-44e2-b72a-08aeb8b97ac4" (UID: "deefee78-9e75-44e2-b72a-08aeb8b97ac4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:06:17 crc kubenswrapper[4934]: I1227 08:06:17.952828 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deefee78-9e75-44e2-b72a-08aeb8b97ac4-kube-api-access-55fc2" (OuterVolumeSpecName: "kube-api-access-55fc2") pod "deefee78-9e75-44e2-b72a-08aeb8b97ac4" (UID: "deefee78-9e75-44e2-b72a-08aeb8b97ac4"). InnerVolumeSpecName "kube-api-access-55fc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:06:17 crc kubenswrapper[4934]: I1227 08:06:17.956864 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-scripts" (OuterVolumeSpecName: "scripts") pod "deefee78-9e75-44e2-b72a-08aeb8b97ac4" (UID: "deefee78-9e75-44e2-b72a-08aeb8b97ac4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:17 crc kubenswrapper[4934]: I1227 08:06:17.993927 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "deefee78-9e75-44e2-b72a-08aeb8b97ac4" (UID: "deefee78-9e75-44e2-b72a-08aeb8b97ac4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.033301 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "deefee78-9e75-44e2-b72a-08aeb8b97ac4" (UID: "deefee78-9e75-44e2-b72a-08aeb8b97ac4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.051893 4934 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.051926 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.051961 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55fc2\" (UniqueName: \"kubernetes.io/projected/deefee78-9e75-44e2-b72a-08aeb8b97ac4-kube-api-access-55fc2\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.051979 4934 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deefee78-9e75-44e2-b72a-08aeb8b97ac4-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.051990 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.052001 4934 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/deefee78-9e75-44e2-b72a-08aeb8b97ac4-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.084426 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-config-data" (OuterVolumeSpecName: "config-data") pod "deefee78-9e75-44e2-b72a-08aeb8b97ac4" (UID: "deefee78-9e75-44e2-b72a-08aeb8b97ac4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.106282 4934 generic.go:334] "Generic (PLEG): container finished" podID="deefee78-9e75-44e2-b72a-08aeb8b97ac4" containerID="62d0acb0da855b7313ed569eab92f4246077da92cd6e3596cc95bd0942f77b98" exitCode=137 Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.106321 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deefee78-9e75-44e2-b72a-08aeb8b97ac4","Type":"ContainerDied","Data":"62d0acb0da855b7313ed569eab92f4246077da92cd6e3596cc95bd0942f77b98"} Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.106360 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.106383 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"deefee78-9e75-44e2-b72a-08aeb8b97ac4","Type":"ContainerDied","Data":"30c22cd6fb03d572171d77615c3621800414a0c1cae613fb9f5dcd05e6d2a7f4"} Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.106400 4934 scope.go:117] "RemoveContainer" containerID="62d0acb0da855b7313ed569eab92f4246077da92cd6e3596cc95bd0942f77b98" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.134746 4934 scope.go:117] "RemoveContainer" containerID="f6b25f19435e4558d201238c1e37486a77141d8977bbfd3f0a3dfb2249350764" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.146718 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.156143 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/deefee78-9e75-44e2-b72a-08aeb8b97ac4-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.159477 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.167128 4934 scope.go:117] "RemoveContainer" containerID="ff283c40ff7fc3fffc16d3b5341ce187559ba24fdae536c0f89834ef0197cc75" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.180113 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:06:18 crc kubenswrapper[4934]: E1227 08:06:18.180623 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deefee78-9e75-44e2-b72a-08aeb8b97ac4" containerName="ceilometer-notification-agent" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.180644 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="deefee78-9e75-44e2-b72a-08aeb8b97ac4" containerName="ceilometer-notification-agent" Dec 27 08:06:18 crc kubenswrapper[4934]: E1227 08:06:18.180668 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deefee78-9e75-44e2-b72a-08aeb8b97ac4" containerName="ceilometer-central-agent" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.180676 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="deefee78-9e75-44e2-b72a-08aeb8b97ac4" containerName="ceilometer-central-agent" Dec 27 08:06:18 crc kubenswrapper[4934]: E1227 08:06:18.180690 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c8feac2-c2ee-484c-baed-052bd7c0b468" containerName="heat-engine" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.180698 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c8feac2-c2ee-484c-baed-052bd7c0b468" containerName="heat-engine" Dec 27 08:06:18 crc kubenswrapper[4934]: E1227 08:06:18.180721 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deefee78-9e75-44e2-b72a-08aeb8b97ac4" containerName="proxy-httpd" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.180746 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="deefee78-9e75-44e2-b72a-08aeb8b97ac4" containerName="proxy-httpd" Dec 27 08:06:18 crc kubenswrapper[4934]: E1227 08:06:18.180769 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deefee78-9e75-44e2-b72a-08aeb8b97ac4" containerName="sg-core" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.180777 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="deefee78-9e75-44e2-b72a-08aeb8b97ac4" containerName="sg-core" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.181197 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="deefee78-9e75-44e2-b72a-08aeb8b97ac4" containerName="ceilometer-notification-agent" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.181221 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="deefee78-9e75-44e2-b72a-08aeb8b97ac4" containerName="ceilometer-central-agent" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.181240 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c8feac2-c2ee-484c-baed-052bd7c0b468" containerName="heat-engine" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.181263 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="deefee78-9e75-44e2-b72a-08aeb8b97ac4" containerName="proxy-httpd" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.181291 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="deefee78-9e75-44e2-b72a-08aeb8b97ac4" containerName="sg-core" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.221888 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.221978 4934 scope.go:117] "RemoveContainer" containerID="c7aef981d05c3a0478c37d7d09c12e1d2edce81b6b397585f3f3259f03f9b58e" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.222265 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.224802 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.225011 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.258891 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-scripts\") pod \"ceilometer-0\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.258960 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.259035 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-log-httpd\") pod \"ceilometer-0\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.259278 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.259347 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-config-data\") pod \"ceilometer-0\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.259423 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wbzk\" (UniqueName: \"kubernetes.io/projected/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-kube-api-access-4wbzk\") pod \"ceilometer-0\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.259449 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-run-httpd\") pod \"ceilometer-0\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.270968 4934 scope.go:117] "RemoveContainer" containerID="62d0acb0da855b7313ed569eab92f4246077da92cd6e3596cc95bd0942f77b98" Dec 27 08:06:18 crc kubenswrapper[4934]: E1227 08:06:18.271419 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62d0acb0da855b7313ed569eab92f4246077da92cd6e3596cc95bd0942f77b98\": container with ID starting with 62d0acb0da855b7313ed569eab92f4246077da92cd6e3596cc95bd0942f77b98 not found: ID does not exist" containerID="62d0acb0da855b7313ed569eab92f4246077da92cd6e3596cc95bd0942f77b98" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.271463 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62d0acb0da855b7313ed569eab92f4246077da92cd6e3596cc95bd0942f77b98"} err="failed to get container status \"62d0acb0da855b7313ed569eab92f4246077da92cd6e3596cc95bd0942f77b98\": rpc error: code = NotFound desc = could not find container \"62d0acb0da855b7313ed569eab92f4246077da92cd6e3596cc95bd0942f77b98\": container with ID starting with 62d0acb0da855b7313ed569eab92f4246077da92cd6e3596cc95bd0942f77b98 not found: ID does not exist" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.271489 4934 scope.go:117] "RemoveContainer" containerID="f6b25f19435e4558d201238c1e37486a77141d8977bbfd3f0a3dfb2249350764" Dec 27 08:06:18 crc kubenswrapper[4934]: E1227 08:06:18.272271 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6b25f19435e4558d201238c1e37486a77141d8977bbfd3f0a3dfb2249350764\": container with ID starting with f6b25f19435e4558d201238c1e37486a77141d8977bbfd3f0a3dfb2249350764 not found: ID does not exist" containerID="f6b25f19435e4558d201238c1e37486a77141d8977bbfd3f0a3dfb2249350764" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.272298 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6b25f19435e4558d201238c1e37486a77141d8977bbfd3f0a3dfb2249350764"} err="failed to get container status \"f6b25f19435e4558d201238c1e37486a77141d8977bbfd3f0a3dfb2249350764\": rpc error: code = NotFound desc = could not find container \"f6b25f19435e4558d201238c1e37486a77141d8977bbfd3f0a3dfb2249350764\": container with ID starting with f6b25f19435e4558d201238c1e37486a77141d8977bbfd3f0a3dfb2249350764 not found: ID does not exist" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.272319 4934 scope.go:117] "RemoveContainer" containerID="ff283c40ff7fc3fffc16d3b5341ce187559ba24fdae536c0f89834ef0197cc75" Dec 27 08:06:18 crc kubenswrapper[4934]: E1227 08:06:18.272612 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff283c40ff7fc3fffc16d3b5341ce187559ba24fdae536c0f89834ef0197cc75\": container with ID starting with ff283c40ff7fc3fffc16d3b5341ce187559ba24fdae536c0f89834ef0197cc75 not found: ID does not exist" containerID="ff283c40ff7fc3fffc16d3b5341ce187559ba24fdae536c0f89834ef0197cc75" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.272639 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff283c40ff7fc3fffc16d3b5341ce187559ba24fdae536c0f89834ef0197cc75"} err="failed to get container status \"ff283c40ff7fc3fffc16d3b5341ce187559ba24fdae536c0f89834ef0197cc75\": rpc error: code = NotFound desc = could not find container \"ff283c40ff7fc3fffc16d3b5341ce187559ba24fdae536c0f89834ef0197cc75\": container with ID starting with ff283c40ff7fc3fffc16d3b5341ce187559ba24fdae536c0f89834ef0197cc75 not found: ID does not exist" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.272658 4934 scope.go:117] "RemoveContainer" containerID="c7aef981d05c3a0478c37d7d09c12e1d2edce81b6b397585f3f3259f03f9b58e" Dec 27 08:06:18 crc kubenswrapper[4934]: E1227 08:06:18.273131 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7aef981d05c3a0478c37d7d09c12e1d2edce81b6b397585f3f3259f03f9b58e\": container with ID starting with c7aef981d05c3a0478c37d7d09c12e1d2edce81b6b397585f3f3259f03f9b58e not found: ID does not exist" containerID="c7aef981d05c3a0478c37d7d09c12e1d2edce81b6b397585f3f3259f03f9b58e" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.273157 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7aef981d05c3a0478c37d7d09c12e1d2edce81b6b397585f3f3259f03f9b58e"} err="failed to get container status \"c7aef981d05c3a0478c37d7d09c12e1d2edce81b6b397585f3f3259f03f9b58e\": rpc error: code = NotFound desc = could not find container \"c7aef981d05c3a0478c37d7d09c12e1d2edce81b6b397585f3f3259f03f9b58e\": container with ID starting with c7aef981d05c3a0478c37d7d09c12e1d2edce81b6b397585f3f3259f03f9b58e not found: ID does not exist" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.361255 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-scripts\") pod \"ceilometer-0\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.361294 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.361321 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-log-httpd\") pod \"ceilometer-0\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.361420 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.361444 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-config-data\") pod \"ceilometer-0\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.361487 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wbzk\" (UniqueName: \"kubernetes.io/projected/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-kube-api-access-4wbzk\") pod \"ceilometer-0\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.361503 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-run-httpd\") pod \"ceilometer-0\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.361827 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-run-httpd\") pod \"ceilometer-0\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.362770 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-log-httpd\") pod \"ceilometer-0\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.366010 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-scripts\") pod \"ceilometer-0\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.366055 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.366287 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-config-data\") pod \"ceilometer-0\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.367110 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.378294 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wbzk\" (UniqueName: \"kubernetes.io/projected/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-kube-api-access-4wbzk\") pod \"ceilometer-0\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " pod="openstack/ceilometer-0" Dec 27 08:06:18 crc kubenswrapper[4934]: I1227 08:06:18.546389 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:06:19 crc kubenswrapper[4934]: W1227 08:06:19.026241 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83835fa6_aa0b_41c8_8f0f_80f99eff15bc.slice/crio-5089d9a7bf02a69febf239b6e2492b0de7aa219b90bb04ceaeca6063cdae72c1 WatchSource:0}: Error finding container 5089d9a7bf02a69febf239b6e2492b0de7aa219b90bb04ceaeca6063cdae72c1: Status 404 returned error can't find the container with id 5089d9a7bf02a69febf239b6e2492b0de7aa219b90bb04ceaeca6063cdae72c1 Dec 27 08:06:19 crc kubenswrapper[4934]: I1227 08:06:19.027099 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:06:19 crc kubenswrapper[4934]: I1227 08:06:19.119484 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83835fa6-aa0b-41c8-8f0f-80f99eff15bc","Type":"ContainerStarted","Data":"5089d9a7bf02a69febf239b6e2492b0de7aa219b90bb04ceaeca6063cdae72c1"} Dec 27 08:06:19 crc kubenswrapper[4934]: I1227 08:06:19.486783 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="deefee78-9e75-44e2-b72a-08aeb8b97ac4" path="/var/lib/kubelet/pods/deefee78-9e75-44e2-b72a-08aeb8b97ac4/volumes" Dec 27 08:06:20 crc kubenswrapper[4934]: I1227 08:06:20.133005 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83835fa6-aa0b-41c8-8f0f-80f99eff15bc","Type":"ContainerStarted","Data":"f3268d5d9897df5fb6a931941c253775a7fb4c9306b325f748b368a8bd0fd630"} Dec 27 08:06:20 crc kubenswrapper[4934]: I1227 08:06:20.629764 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:06:21 crc kubenswrapper[4934]: I1227 08:06:21.154952 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83835fa6-aa0b-41c8-8f0f-80f99eff15bc","Type":"ContainerStarted","Data":"6148c29b562ed06b874838be5b7e70da3b8414e79db632e77dac592672a699f7"} Dec 27 08:06:22 crc kubenswrapper[4934]: I1227 08:06:22.170648 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83835fa6-aa0b-41c8-8f0f-80f99eff15bc","Type":"ContainerStarted","Data":"f412beda5ee70f0c375ad18760eadeee11e571ab3fca827ce11b3ccd72940360"} Dec 27 08:06:23 crc kubenswrapper[4934]: I1227 08:06:23.184131 4934 generic.go:334] "Generic (PLEG): container finished" podID="85e47792-15be-4b6b-8a8e-8c05c4ae4605" containerID="a25540b169ccd286d2daa333158ae981f01dc148e21953b9855b096d30040eb1" exitCode=0 Dec 27 08:06:23 crc kubenswrapper[4934]: I1227 08:06:23.184215 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-cm8ph" event={"ID":"85e47792-15be-4b6b-8a8e-8c05c4ae4605","Type":"ContainerDied","Data":"a25540b169ccd286d2daa333158ae981f01dc148e21953b9855b096d30040eb1"} Dec 27 08:06:23 crc kubenswrapper[4934]: I1227 08:06:23.188047 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83835fa6-aa0b-41c8-8f0f-80f99eff15bc","Type":"ContainerStarted","Data":"d7bf93acbedd6de2c9252cf24be1c90e74e6fa2dcbc81dbfd166f01c060c9684"} Dec 27 08:06:23 crc kubenswrapper[4934]: I1227 08:06:23.188282 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 27 08:06:23 crc kubenswrapper[4934]: I1227 08:06:23.188328 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="83835fa6-aa0b-41c8-8f0f-80f99eff15bc" containerName="ceilometer-notification-agent" containerID="cri-o://6148c29b562ed06b874838be5b7e70da3b8414e79db632e77dac592672a699f7" gracePeriod=30 Dec 27 08:06:23 crc kubenswrapper[4934]: I1227 08:06:23.188330 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="83835fa6-aa0b-41c8-8f0f-80f99eff15bc" containerName="sg-core" containerID="cri-o://f412beda5ee70f0c375ad18760eadeee11e571ab3fca827ce11b3ccd72940360" gracePeriod=30 Dec 27 08:06:23 crc kubenswrapper[4934]: I1227 08:06:23.188323 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="83835fa6-aa0b-41c8-8f0f-80f99eff15bc" containerName="ceilometer-central-agent" containerID="cri-o://f3268d5d9897df5fb6a931941c253775a7fb4c9306b325f748b368a8bd0fd630" gracePeriod=30 Dec 27 08:06:23 crc kubenswrapper[4934]: I1227 08:06:23.188484 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="83835fa6-aa0b-41c8-8f0f-80f99eff15bc" containerName="proxy-httpd" containerID="cri-o://d7bf93acbedd6de2c9252cf24be1c90e74e6fa2dcbc81dbfd166f01c060c9684" gracePeriod=30 Dec 27 08:06:24 crc kubenswrapper[4934]: I1227 08:06:24.199826 4934 generic.go:334] "Generic (PLEG): container finished" podID="83835fa6-aa0b-41c8-8f0f-80f99eff15bc" containerID="d7bf93acbedd6de2c9252cf24be1c90e74e6fa2dcbc81dbfd166f01c060c9684" exitCode=0 Dec 27 08:06:24 crc kubenswrapper[4934]: I1227 08:06:24.200177 4934 generic.go:334] "Generic (PLEG): container finished" podID="83835fa6-aa0b-41c8-8f0f-80f99eff15bc" containerID="f412beda5ee70f0c375ad18760eadeee11e571ab3fca827ce11b3ccd72940360" exitCode=2 Dec 27 08:06:24 crc kubenswrapper[4934]: I1227 08:06:24.200188 4934 generic.go:334] "Generic (PLEG): container finished" podID="83835fa6-aa0b-41c8-8f0f-80f99eff15bc" containerID="6148c29b562ed06b874838be5b7e70da3b8414e79db632e77dac592672a699f7" exitCode=0 Dec 27 08:06:24 crc kubenswrapper[4934]: I1227 08:06:24.199879 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83835fa6-aa0b-41c8-8f0f-80f99eff15bc","Type":"ContainerDied","Data":"d7bf93acbedd6de2c9252cf24be1c90e74e6fa2dcbc81dbfd166f01c060c9684"} Dec 27 08:06:24 crc kubenswrapper[4934]: I1227 08:06:24.200319 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83835fa6-aa0b-41c8-8f0f-80f99eff15bc","Type":"ContainerDied","Data":"f412beda5ee70f0c375ad18760eadeee11e571ab3fca827ce11b3ccd72940360"} Dec 27 08:06:24 crc kubenswrapper[4934]: I1227 08:06:24.200349 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83835fa6-aa0b-41c8-8f0f-80f99eff15bc","Type":"ContainerDied","Data":"6148c29b562ed06b874838be5b7e70da3b8414e79db632e77dac592672a699f7"} Dec 27 08:06:24 crc kubenswrapper[4934]: I1227 08:06:24.590326 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-cm8ph" Dec 27 08:06:24 crc kubenswrapper[4934]: I1227 08:06:24.609628 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.948219534 podStartE2EDuration="6.609612913s" podCreationTimestamp="2025-12-27 08:06:18 +0000 UTC" firstStartedPulling="2025-12-27 08:06:19.028490689 +0000 UTC m=+1439.848931283" lastFinishedPulling="2025-12-27 08:06:22.689884048 +0000 UTC m=+1443.510324662" observedRunningTime="2025-12-27 08:06:23.264511103 +0000 UTC m=+1444.084951747" watchObservedRunningTime="2025-12-27 08:06:24.609612913 +0000 UTC m=+1445.430053507" Dec 27 08:06:24 crc kubenswrapper[4934]: I1227 08:06:24.718489 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85e47792-15be-4b6b-8a8e-8c05c4ae4605-combined-ca-bundle\") pod \"85e47792-15be-4b6b-8a8e-8c05c4ae4605\" (UID: \"85e47792-15be-4b6b-8a8e-8c05c4ae4605\") " Dec 27 08:06:24 crc kubenswrapper[4934]: I1227 08:06:24.718614 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrxd9\" (UniqueName: \"kubernetes.io/projected/85e47792-15be-4b6b-8a8e-8c05c4ae4605-kube-api-access-hrxd9\") pod \"85e47792-15be-4b6b-8a8e-8c05c4ae4605\" (UID: \"85e47792-15be-4b6b-8a8e-8c05c4ae4605\") " Dec 27 08:06:24 crc kubenswrapper[4934]: I1227 08:06:24.718686 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85e47792-15be-4b6b-8a8e-8c05c4ae4605-config-data\") pod \"85e47792-15be-4b6b-8a8e-8c05c4ae4605\" (UID: \"85e47792-15be-4b6b-8a8e-8c05c4ae4605\") " Dec 27 08:06:24 crc kubenswrapper[4934]: I1227 08:06:24.718784 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85e47792-15be-4b6b-8a8e-8c05c4ae4605-scripts\") pod \"85e47792-15be-4b6b-8a8e-8c05c4ae4605\" (UID: \"85e47792-15be-4b6b-8a8e-8c05c4ae4605\") " Dec 27 08:06:24 crc kubenswrapper[4934]: I1227 08:06:24.737138 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85e47792-15be-4b6b-8a8e-8c05c4ae4605-kube-api-access-hrxd9" (OuterVolumeSpecName: "kube-api-access-hrxd9") pod "85e47792-15be-4b6b-8a8e-8c05c4ae4605" (UID: "85e47792-15be-4b6b-8a8e-8c05c4ae4605"). InnerVolumeSpecName "kube-api-access-hrxd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:06:24 crc kubenswrapper[4934]: I1227 08:06:24.737285 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85e47792-15be-4b6b-8a8e-8c05c4ae4605-scripts" (OuterVolumeSpecName: "scripts") pod "85e47792-15be-4b6b-8a8e-8c05c4ae4605" (UID: "85e47792-15be-4b6b-8a8e-8c05c4ae4605"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:24 crc kubenswrapper[4934]: I1227 08:06:24.754866 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85e47792-15be-4b6b-8a8e-8c05c4ae4605-config-data" (OuterVolumeSpecName: "config-data") pod "85e47792-15be-4b6b-8a8e-8c05c4ae4605" (UID: "85e47792-15be-4b6b-8a8e-8c05c4ae4605"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:24 crc kubenswrapper[4934]: I1227 08:06:24.769942 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85e47792-15be-4b6b-8a8e-8c05c4ae4605-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85e47792-15be-4b6b-8a8e-8c05c4ae4605" (UID: "85e47792-15be-4b6b-8a8e-8c05c4ae4605"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:24 crc kubenswrapper[4934]: I1227 08:06:24.821408 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrxd9\" (UniqueName: \"kubernetes.io/projected/85e47792-15be-4b6b-8a8e-8c05c4ae4605-kube-api-access-hrxd9\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:24 crc kubenswrapper[4934]: I1227 08:06:24.821446 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85e47792-15be-4b6b-8a8e-8c05c4ae4605-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:24 crc kubenswrapper[4934]: I1227 08:06:24.821462 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85e47792-15be-4b6b-8a8e-8c05c4ae4605-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:24 crc kubenswrapper[4934]: I1227 08:06:24.821475 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85e47792-15be-4b6b-8a8e-8c05c4ae4605-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:25 crc kubenswrapper[4934]: I1227 08:06:25.212497 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-cm8ph" event={"ID":"85e47792-15be-4b6b-8a8e-8c05c4ae4605","Type":"ContainerDied","Data":"c7bc93e79f7230a8abc78e7e83b917dbb13c98c0af36e810d676fb4075b8a1d7"} Dec 27 08:06:25 crc kubenswrapper[4934]: I1227 08:06:25.212540 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7bc93e79f7230a8abc78e7e83b917dbb13c98c0af36e810d676fb4075b8a1d7" Dec 27 08:06:25 crc kubenswrapper[4934]: I1227 08:06:25.212552 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-cm8ph" Dec 27 08:06:25 crc kubenswrapper[4934]: I1227 08:06:25.314828 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 27 08:06:25 crc kubenswrapper[4934]: E1227 08:06:25.315406 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85e47792-15be-4b6b-8a8e-8c05c4ae4605" containerName="nova-cell0-conductor-db-sync" Dec 27 08:06:25 crc kubenswrapper[4934]: I1227 08:06:25.315424 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="85e47792-15be-4b6b-8a8e-8c05c4ae4605" containerName="nova-cell0-conductor-db-sync" Dec 27 08:06:25 crc kubenswrapper[4934]: I1227 08:06:25.315655 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="85e47792-15be-4b6b-8a8e-8c05c4ae4605" containerName="nova-cell0-conductor-db-sync" Dec 27 08:06:25 crc kubenswrapper[4934]: I1227 08:06:25.316523 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 27 08:06:25 crc kubenswrapper[4934]: I1227 08:06:25.318904 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 27 08:06:25 crc kubenswrapper[4934]: I1227 08:06:25.319001 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-tlgks" Dec 27 08:06:25 crc kubenswrapper[4934]: I1227 08:06:25.343644 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 27 08:06:25 crc kubenswrapper[4934]: I1227 08:06:25.434833 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvcmw\" (UniqueName: \"kubernetes.io/projected/4507dd01-8c5e-482f-8881-6e0c70679bec-kube-api-access-wvcmw\") pod \"nova-cell0-conductor-0\" (UID: \"4507dd01-8c5e-482f-8881-6e0c70679bec\") " pod="openstack/nova-cell0-conductor-0" Dec 27 08:06:25 crc kubenswrapper[4934]: I1227 08:06:25.435280 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4507dd01-8c5e-482f-8881-6e0c70679bec-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4507dd01-8c5e-482f-8881-6e0c70679bec\") " pod="openstack/nova-cell0-conductor-0" Dec 27 08:06:25 crc kubenswrapper[4934]: I1227 08:06:25.435336 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4507dd01-8c5e-482f-8881-6e0c70679bec-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4507dd01-8c5e-482f-8881-6e0c70679bec\") " pod="openstack/nova-cell0-conductor-0" Dec 27 08:06:25 crc kubenswrapper[4934]: I1227 08:06:25.538072 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4507dd01-8c5e-482f-8881-6e0c70679bec-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4507dd01-8c5e-482f-8881-6e0c70679bec\") " pod="openstack/nova-cell0-conductor-0" Dec 27 08:06:25 crc kubenswrapper[4934]: I1227 08:06:25.538342 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4507dd01-8c5e-482f-8881-6e0c70679bec-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4507dd01-8c5e-482f-8881-6e0c70679bec\") " pod="openstack/nova-cell0-conductor-0" Dec 27 08:06:25 crc kubenswrapper[4934]: I1227 08:06:25.538480 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvcmw\" (UniqueName: \"kubernetes.io/projected/4507dd01-8c5e-482f-8881-6e0c70679bec-kube-api-access-wvcmw\") pod \"nova-cell0-conductor-0\" (UID: \"4507dd01-8c5e-482f-8881-6e0c70679bec\") " pod="openstack/nova-cell0-conductor-0" Dec 27 08:06:25 crc kubenswrapper[4934]: I1227 08:06:25.542159 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4507dd01-8c5e-482f-8881-6e0c70679bec-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4507dd01-8c5e-482f-8881-6e0c70679bec\") " pod="openstack/nova-cell0-conductor-0" Dec 27 08:06:25 crc kubenswrapper[4934]: I1227 08:06:25.543783 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4507dd01-8c5e-482f-8881-6e0c70679bec-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4507dd01-8c5e-482f-8881-6e0c70679bec\") " pod="openstack/nova-cell0-conductor-0" Dec 27 08:06:25 crc kubenswrapper[4934]: I1227 08:06:25.554660 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvcmw\" (UniqueName: \"kubernetes.io/projected/4507dd01-8c5e-482f-8881-6e0c70679bec-kube-api-access-wvcmw\") pod \"nova-cell0-conductor-0\" (UID: \"4507dd01-8c5e-482f-8881-6e0c70679bec\") " pod="openstack/nova-cell0-conductor-0" Dec 27 08:06:25 crc kubenswrapper[4934]: I1227 08:06:25.631371 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 27 08:06:26 crc kubenswrapper[4934]: I1227 08:06:26.175931 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 27 08:06:26 crc kubenswrapper[4934]: W1227 08:06:26.201577 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4507dd01_8c5e_482f_8881_6e0c70679bec.slice/crio-d585f6cae91f65969a2b8a36e689c53a0637d2526962abd08c8acf18e2668a8d WatchSource:0}: Error finding container d585f6cae91f65969a2b8a36e689c53a0637d2526962abd08c8acf18e2668a8d: Status 404 returned error can't find the container with id d585f6cae91f65969a2b8a36e689c53a0637d2526962abd08c8acf18e2668a8d Dec 27 08:06:26 crc kubenswrapper[4934]: I1227 08:06:26.231634 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4507dd01-8c5e-482f-8881-6e0c70679bec","Type":"ContainerStarted","Data":"d585f6cae91f65969a2b8a36e689c53a0637d2526962abd08c8acf18e2668a8d"} Dec 27 08:06:27 crc kubenswrapper[4934]: I1227 08:06:27.245644 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4507dd01-8c5e-482f-8881-6e0c70679bec","Type":"ContainerStarted","Data":"9ca76d61d7ada5ef4fc1e326c21411366fc9cfa4d1f711c33fba4127b2917d75"} Dec 27 08:06:27 crc kubenswrapper[4934]: I1227 08:06:27.246346 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 27 08:06:27 crc kubenswrapper[4934]: I1227 08:06:27.270356 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.2703371949999998 podStartE2EDuration="2.270337195s" podCreationTimestamp="2025-12-27 08:06:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:06:27.261688312 +0000 UTC m=+1448.082128906" watchObservedRunningTime="2025-12-27 08:06:27.270337195 +0000 UTC m=+1448.090777789" Dec 27 08:06:30 crc kubenswrapper[4934]: I1227 08:06:30.846831 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:06:30 crc kubenswrapper[4934]: I1227 08:06:30.975605 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wbzk\" (UniqueName: \"kubernetes.io/projected/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-kube-api-access-4wbzk\") pod \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " Dec 27 08:06:30 crc kubenswrapper[4934]: I1227 08:06:30.975670 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-config-data\") pod \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " Dec 27 08:06:30 crc kubenswrapper[4934]: I1227 08:06:30.975765 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-log-httpd\") pod \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " Dec 27 08:06:30 crc kubenswrapper[4934]: I1227 08:06:30.975827 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-scripts\") pod \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " Dec 27 08:06:30 crc kubenswrapper[4934]: I1227 08:06:30.975882 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-combined-ca-bundle\") pod \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " Dec 27 08:06:30 crc kubenswrapper[4934]: I1227 08:06:30.975962 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-sg-core-conf-yaml\") pod \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " Dec 27 08:06:30 crc kubenswrapper[4934]: I1227 08:06:30.976100 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-run-httpd\") pod \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\" (UID: \"83835fa6-aa0b-41c8-8f0f-80f99eff15bc\") " Dec 27 08:06:30 crc kubenswrapper[4934]: I1227 08:06:30.977279 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "83835fa6-aa0b-41c8-8f0f-80f99eff15bc" (UID: "83835fa6-aa0b-41c8-8f0f-80f99eff15bc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:06:30 crc kubenswrapper[4934]: I1227 08:06:30.980181 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "83835fa6-aa0b-41c8-8f0f-80f99eff15bc" (UID: "83835fa6-aa0b-41c8-8f0f-80f99eff15bc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:06:30 crc kubenswrapper[4934]: I1227 08:06:30.984182 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-kube-api-access-4wbzk" (OuterVolumeSpecName: "kube-api-access-4wbzk") pod "83835fa6-aa0b-41c8-8f0f-80f99eff15bc" (UID: "83835fa6-aa0b-41c8-8f0f-80f99eff15bc"). InnerVolumeSpecName "kube-api-access-4wbzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:06:30 crc kubenswrapper[4934]: I1227 08:06:30.991143 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-scripts" (OuterVolumeSpecName: "scripts") pod "83835fa6-aa0b-41c8-8f0f-80f99eff15bc" (UID: "83835fa6-aa0b-41c8-8f0f-80f99eff15bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.030282 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "83835fa6-aa0b-41c8-8f0f-80f99eff15bc" (UID: "83835fa6-aa0b-41c8-8f0f-80f99eff15bc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.077372 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83835fa6-aa0b-41c8-8f0f-80f99eff15bc" (UID: "83835fa6-aa0b-41c8-8f0f-80f99eff15bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.078586 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.078745 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.078755 4934 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.078763 4934 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.078773 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wbzk\" (UniqueName: \"kubernetes.io/projected/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-kube-api-access-4wbzk\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.078782 4934 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.100880 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-config-data" (OuterVolumeSpecName: "config-data") pod "83835fa6-aa0b-41c8-8f0f-80f99eff15bc" (UID: "83835fa6-aa0b-41c8-8f0f-80f99eff15bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.180817 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83835fa6-aa0b-41c8-8f0f-80f99eff15bc-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.307963 4934 generic.go:334] "Generic (PLEG): container finished" podID="83835fa6-aa0b-41c8-8f0f-80f99eff15bc" containerID="f3268d5d9897df5fb6a931941c253775a7fb4c9306b325f748b368a8bd0fd630" exitCode=0 Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.308008 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83835fa6-aa0b-41c8-8f0f-80f99eff15bc","Type":"ContainerDied","Data":"f3268d5d9897df5fb6a931941c253775a7fb4c9306b325f748b368a8bd0fd630"} Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.308040 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"83835fa6-aa0b-41c8-8f0f-80f99eff15bc","Type":"ContainerDied","Data":"5089d9a7bf02a69febf239b6e2492b0de7aa219b90bb04ceaeca6063cdae72c1"} Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.308061 4934 scope.go:117] "RemoveContainer" containerID="d7bf93acbedd6de2c9252cf24be1c90e74e6fa2dcbc81dbfd166f01c060c9684" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.308169 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.368301 4934 scope.go:117] "RemoveContainer" containerID="f412beda5ee70f0c375ad18760eadeee11e571ab3fca827ce11b3ccd72940360" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.425005 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.433180 4934 scope.go:117] "RemoveContainer" containerID="6148c29b562ed06b874838be5b7e70da3b8414e79db632e77dac592672a699f7" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.460266 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.481063 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83835fa6-aa0b-41c8-8f0f-80f99eff15bc" path="/var/lib/kubelet/pods/83835fa6-aa0b-41c8-8f0f-80f99eff15bc/volumes" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.482169 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:06:31 crc kubenswrapper[4934]: E1227 08:06:31.482594 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83835fa6-aa0b-41c8-8f0f-80f99eff15bc" containerName="ceilometer-central-agent" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.482613 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="83835fa6-aa0b-41c8-8f0f-80f99eff15bc" containerName="ceilometer-central-agent" Dec 27 08:06:31 crc kubenswrapper[4934]: E1227 08:06:31.482635 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83835fa6-aa0b-41c8-8f0f-80f99eff15bc" containerName="proxy-httpd" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.482644 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="83835fa6-aa0b-41c8-8f0f-80f99eff15bc" containerName="proxy-httpd" Dec 27 08:06:31 crc kubenswrapper[4934]: E1227 08:06:31.482674 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83835fa6-aa0b-41c8-8f0f-80f99eff15bc" containerName="sg-core" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.482682 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="83835fa6-aa0b-41c8-8f0f-80f99eff15bc" containerName="sg-core" Dec 27 08:06:31 crc kubenswrapper[4934]: E1227 08:06:31.482719 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83835fa6-aa0b-41c8-8f0f-80f99eff15bc" containerName="ceilometer-notification-agent" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.482728 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="83835fa6-aa0b-41c8-8f0f-80f99eff15bc" containerName="ceilometer-notification-agent" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.482971 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="83835fa6-aa0b-41c8-8f0f-80f99eff15bc" containerName="sg-core" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.482998 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="83835fa6-aa0b-41c8-8f0f-80f99eff15bc" containerName="ceilometer-notification-agent" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.483015 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="83835fa6-aa0b-41c8-8f0f-80f99eff15bc" containerName="proxy-httpd" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.483027 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="83835fa6-aa0b-41c8-8f0f-80f99eff15bc" containerName="ceilometer-central-agent" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.486365 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.488014 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.488634 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.488783 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.523961 4934 scope.go:117] "RemoveContainer" containerID="f3268d5d9897df5fb6a931941c253775a7fb4c9306b325f748b368a8bd0fd630" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.546273 4934 scope.go:117] "RemoveContainer" containerID="d7bf93acbedd6de2c9252cf24be1c90e74e6fa2dcbc81dbfd166f01c060c9684" Dec 27 08:06:31 crc kubenswrapper[4934]: E1227 08:06:31.546723 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7bf93acbedd6de2c9252cf24be1c90e74e6fa2dcbc81dbfd166f01c060c9684\": container with ID starting with d7bf93acbedd6de2c9252cf24be1c90e74e6fa2dcbc81dbfd166f01c060c9684 not found: ID does not exist" containerID="d7bf93acbedd6de2c9252cf24be1c90e74e6fa2dcbc81dbfd166f01c060c9684" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.546772 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7bf93acbedd6de2c9252cf24be1c90e74e6fa2dcbc81dbfd166f01c060c9684"} err="failed to get container status \"d7bf93acbedd6de2c9252cf24be1c90e74e6fa2dcbc81dbfd166f01c060c9684\": rpc error: code = NotFound desc = could not find container \"d7bf93acbedd6de2c9252cf24be1c90e74e6fa2dcbc81dbfd166f01c060c9684\": container with ID starting with d7bf93acbedd6de2c9252cf24be1c90e74e6fa2dcbc81dbfd166f01c060c9684 not found: ID does not exist" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.546806 4934 scope.go:117] "RemoveContainer" containerID="f412beda5ee70f0c375ad18760eadeee11e571ab3fca827ce11b3ccd72940360" Dec 27 08:06:31 crc kubenswrapper[4934]: E1227 08:06:31.547141 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f412beda5ee70f0c375ad18760eadeee11e571ab3fca827ce11b3ccd72940360\": container with ID starting with f412beda5ee70f0c375ad18760eadeee11e571ab3fca827ce11b3ccd72940360 not found: ID does not exist" containerID="f412beda5ee70f0c375ad18760eadeee11e571ab3fca827ce11b3ccd72940360" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.547174 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f412beda5ee70f0c375ad18760eadeee11e571ab3fca827ce11b3ccd72940360"} err="failed to get container status \"f412beda5ee70f0c375ad18760eadeee11e571ab3fca827ce11b3ccd72940360\": rpc error: code = NotFound desc = could not find container \"f412beda5ee70f0c375ad18760eadeee11e571ab3fca827ce11b3ccd72940360\": container with ID starting with f412beda5ee70f0c375ad18760eadeee11e571ab3fca827ce11b3ccd72940360 not found: ID does not exist" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.547201 4934 scope.go:117] "RemoveContainer" containerID="6148c29b562ed06b874838be5b7e70da3b8414e79db632e77dac592672a699f7" Dec 27 08:06:31 crc kubenswrapper[4934]: E1227 08:06:31.547475 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6148c29b562ed06b874838be5b7e70da3b8414e79db632e77dac592672a699f7\": container with ID starting with 6148c29b562ed06b874838be5b7e70da3b8414e79db632e77dac592672a699f7 not found: ID does not exist" containerID="6148c29b562ed06b874838be5b7e70da3b8414e79db632e77dac592672a699f7" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.547531 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6148c29b562ed06b874838be5b7e70da3b8414e79db632e77dac592672a699f7"} err="failed to get container status \"6148c29b562ed06b874838be5b7e70da3b8414e79db632e77dac592672a699f7\": rpc error: code = NotFound desc = could not find container \"6148c29b562ed06b874838be5b7e70da3b8414e79db632e77dac592672a699f7\": container with ID starting with 6148c29b562ed06b874838be5b7e70da3b8414e79db632e77dac592672a699f7 not found: ID does not exist" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.547572 4934 scope.go:117] "RemoveContainer" containerID="f3268d5d9897df5fb6a931941c253775a7fb4c9306b325f748b368a8bd0fd630" Dec 27 08:06:31 crc kubenswrapper[4934]: E1227 08:06:31.547844 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3268d5d9897df5fb6a931941c253775a7fb4c9306b325f748b368a8bd0fd630\": container with ID starting with f3268d5d9897df5fb6a931941c253775a7fb4c9306b325f748b368a8bd0fd630 not found: ID does not exist" containerID="f3268d5d9897df5fb6a931941c253775a7fb4c9306b325f748b368a8bd0fd630" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.547872 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3268d5d9897df5fb6a931941c253775a7fb4c9306b325f748b368a8bd0fd630"} err="failed to get container status \"f3268d5d9897df5fb6a931941c253775a7fb4c9306b325f748b368a8bd0fd630\": rpc error: code = NotFound desc = could not find container \"f3268d5d9897df5fb6a931941c253775a7fb4c9306b325f748b368a8bd0fd630\": container with ID starting with f3268d5d9897df5fb6a931941c253775a7fb4c9306b325f748b368a8bd0fd630 not found: ID does not exist" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.601756 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.602153 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.602283 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/600b5343-a21b-4c61-94b4-54678000d8de-log-httpd\") pod \"ceilometer-0\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.602348 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/600b5343-a21b-4c61-94b4-54678000d8de-run-httpd\") pod \"ceilometer-0\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.602565 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-config-data\") pod \"ceilometer-0\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.602636 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-scripts\") pod \"ceilometer-0\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.602693 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nhk4\" (UniqueName: \"kubernetes.io/projected/600b5343-a21b-4c61-94b4-54678000d8de-kube-api-access-7nhk4\") pod \"ceilometer-0\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.705221 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-config-data\") pod \"ceilometer-0\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.705295 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-scripts\") pod \"ceilometer-0\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.705330 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nhk4\" (UniqueName: \"kubernetes.io/projected/600b5343-a21b-4c61-94b4-54678000d8de-kube-api-access-7nhk4\") pod \"ceilometer-0\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.705456 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.705540 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.705577 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/600b5343-a21b-4c61-94b4-54678000d8de-log-httpd\") pod \"ceilometer-0\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.705625 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/600b5343-a21b-4c61-94b4-54678000d8de-run-httpd\") pod \"ceilometer-0\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.706194 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/600b5343-a21b-4c61-94b4-54678000d8de-run-httpd\") pod \"ceilometer-0\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.706293 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/600b5343-a21b-4c61-94b4-54678000d8de-log-httpd\") pod \"ceilometer-0\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.709653 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.710123 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-scripts\") pod \"ceilometer-0\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.710177 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-config-data\") pod \"ceilometer-0\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.720662 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.745950 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nhk4\" (UniqueName: \"kubernetes.io/projected/600b5343-a21b-4c61-94b4-54678000d8de-kube-api-access-7nhk4\") pod \"ceilometer-0\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " pod="openstack/ceilometer-0" Dec 27 08:06:31 crc kubenswrapper[4934]: I1227 08:06:31.819129 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:06:32 crc kubenswrapper[4934]: I1227 08:06:32.335323 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:06:33 crc kubenswrapper[4934]: I1227 08:06:33.346726 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"600b5343-a21b-4c61-94b4-54678000d8de","Type":"ContainerStarted","Data":"6dcda01dab092e5244c68ae57ba68aabf16cff612d634574698d3e67e2504305"} Dec 27 08:06:33 crc kubenswrapper[4934]: I1227 08:06:33.347129 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"600b5343-a21b-4c61-94b4-54678000d8de","Type":"ContainerStarted","Data":"3d1805306412f0394ea8cde2cf554718742f81a85edf74294e9f4762e1b653c0"} Dec 27 08:06:33 crc kubenswrapper[4934]: I1227 08:06:33.733772 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-4vl7v"] Dec 27 08:06:33 crc kubenswrapper[4934]: I1227 08:06:33.735314 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-4vl7v" Dec 27 08:06:33 crc kubenswrapper[4934]: I1227 08:06:33.763672 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-4vl7v"] Dec 27 08:06:33 crc kubenswrapper[4934]: I1227 08:06:33.858399 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l76wt\" (UniqueName: \"kubernetes.io/projected/38941274-cf93-4985-9e25-f9aad3dcb394-kube-api-access-l76wt\") pod \"aodh-db-create-4vl7v\" (UID: \"38941274-cf93-4985-9e25-f9aad3dcb394\") " pod="openstack/aodh-db-create-4vl7v" Dec 27 08:06:33 crc kubenswrapper[4934]: I1227 08:06:33.960937 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l76wt\" (UniqueName: \"kubernetes.io/projected/38941274-cf93-4985-9e25-f9aad3dcb394-kube-api-access-l76wt\") pod \"aodh-db-create-4vl7v\" (UID: \"38941274-cf93-4985-9e25-f9aad3dcb394\") " pod="openstack/aodh-db-create-4vl7v" Dec 27 08:06:33 crc kubenswrapper[4934]: I1227 08:06:33.976493 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l76wt\" (UniqueName: \"kubernetes.io/projected/38941274-cf93-4985-9e25-f9aad3dcb394-kube-api-access-l76wt\") pod \"aodh-db-create-4vl7v\" (UID: \"38941274-cf93-4985-9e25-f9aad3dcb394\") " pod="openstack/aodh-db-create-4vl7v" Dec 27 08:06:34 crc kubenswrapper[4934]: I1227 08:06:34.056449 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-4vl7v" Dec 27 08:06:34 crc kubenswrapper[4934]: I1227 08:06:34.372165 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"600b5343-a21b-4c61-94b4-54678000d8de","Type":"ContainerStarted","Data":"7c39932385049c834e0a7ae64966f692854ef783b5795bf025e9f64f9eafe5eb"} Dec 27 08:06:34 crc kubenswrapper[4934]: W1227 08:06:34.530871 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38941274_cf93_4985_9e25_f9aad3dcb394.slice/crio-c538d5506fb7899223a0e0a79e0333b18d418a2aa998011d8163c109cf8eecb3 WatchSource:0}: Error finding container c538d5506fb7899223a0e0a79e0333b18d418a2aa998011d8163c109cf8eecb3: Status 404 returned error can't find the container with id c538d5506fb7899223a0e0a79e0333b18d418a2aa998011d8163c109cf8eecb3 Dec 27 08:06:34 crc kubenswrapper[4934]: I1227 08:06:34.533648 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-4vl7v"] Dec 27 08:06:35 crc kubenswrapper[4934]: I1227 08:06:35.383121 4934 generic.go:334] "Generic (PLEG): container finished" podID="38941274-cf93-4985-9e25-f9aad3dcb394" containerID="efdee1bf217bc262bb57f92fa39bb1ff916bdd3ee785824c29c4a087d23c6b74" exitCode=0 Dec 27 08:06:35 crc kubenswrapper[4934]: I1227 08:06:35.383179 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-4vl7v" event={"ID":"38941274-cf93-4985-9e25-f9aad3dcb394","Type":"ContainerDied","Data":"efdee1bf217bc262bb57f92fa39bb1ff916bdd3ee785824c29c4a087d23c6b74"} Dec 27 08:06:35 crc kubenswrapper[4934]: I1227 08:06:35.383510 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-4vl7v" event={"ID":"38941274-cf93-4985-9e25-f9aad3dcb394","Type":"ContainerStarted","Data":"c538d5506fb7899223a0e0a79e0333b18d418a2aa998011d8163c109cf8eecb3"} Dec 27 08:06:35 crc kubenswrapper[4934]: I1227 08:06:35.385645 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"600b5343-a21b-4c61-94b4-54678000d8de","Type":"ContainerStarted","Data":"36e64ee6387a3cd1b80d20e2a97954d73739eac60a5f9361a858c4a90b0fdcd5"} Dec 27 08:06:35 crc kubenswrapper[4934]: I1227 08:06:35.667546 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.189210 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-qxkzj"] Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.191295 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qxkzj" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.194376 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.194570 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.206322 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-qxkzj"] Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.319110 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-976zx\" (UniqueName: \"kubernetes.io/projected/94199ff4-bea2-4054-9752-7f9e5cd73d99-kube-api-access-976zx\") pod \"nova-cell0-cell-mapping-qxkzj\" (UID: \"94199ff4-bea2-4054-9752-7f9e5cd73d99\") " pod="openstack/nova-cell0-cell-mapping-qxkzj" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.319277 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94199ff4-bea2-4054-9752-7f9e5cd73d99-config-data\") pod \"nova-cell0-cell-mapping-qxkzj\" (UID: \"94199ff4-bea2-4054-9752-7f9e5cd73d99\") " pod="openstack/nova-cell0-cell-mapping-qxkzj" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.319357 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94199ff4-bea2-4054-9752-7f9e5cd73d99-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qxkzj\" (UID: \"94199ff4-bea2-4054-9752-7f9e5cd73d99\") " pod="openstack/nova-cell0-cell-mapping-qxkzj" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.319390 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94199ff4-bea2-4054-9752-7f9e5cd73d99-scripts\") pod \"nova-cell0-cell-mapping-qxkzj\" (UID: \"94199ff4-bea2-4054-9752-7f9e5cd73d99\") " pod="openstack/nova-cell0-cell-mapping-qxkzj" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.334595 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.336300 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.339206 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.378278 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.421266 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94199ff4-bea2-4054-9752-7f9e5cd73d99-scripts\") pod \"nova-cell0-cell-mapping-qxkzj\" (UID: \"94199ff4-bea2-4054-9752-7f9e5cd73d99\") " pod="openstack/nova-cell0-cell-mapping-qxkzj" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.421327 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-976zx\" (UniqueName: \"kubernetes.io/projected/94199ff4-bea2-4054-9752-7f9e5cd73d99-kube-api-access-976zx\") pod \"nova-cell0-cell-mapping-qxkzj\" (UID: \"94199ff4-bea2-4054-9752-7f9e5cd73d99\") " pod="openstack/nova-cell0-cell-mapping-qxkzj" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.421369 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/794826da-01ec-4ee2-99dd-e5ece92df833-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"794826da-01ec-4ee2-99dd-e5ece92df833\") " pod="openstack/nova-scheduler-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.421406 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nptb\" (UniqueName: \"kubernetes.io/projected/794826da-01ec-4ee2-99dd-e5ece92df833-kube-api-access-7nptb\") pod \"nova-scheduler-0\" (UID: \"794826da-01ec-4ee2-99dd-e5ece92df833\") " pod="openstack/nova-scheduler-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.421446 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/794826da-01ec-4ee2-99dd-e5ece92df833-config-data\") pod \"nova-scheduler-0\" (UID: \"794826da-01ec-4ee2-99dd-e5ece92df833\") " pod="openstack/nova-scheduler-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.421518 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94199ff4-bea2-4054-9752-7f9e5cd73d99-config-data\") pod \"nova-cell0-cell-mapping-qxkzj\" (UID: \"94199ff4-bea2-4054-9752-7f9e5cd73d99\") " pod="openstack/nova-cell0-cell-mapping-qxkzj" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.421577 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94199ff4-bea2-4054-9752-7f9e5cd73d99-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qxkzj\" (UID: \"94199ff4-bea2-4054-9752-7f9e5cd73d99\") " pod="openstack/nova-cell0-cell-mapping-qxkzj" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.449966 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94199ff4-bea2-4054-9752-7f9e5cd73d99-config-data\") pod \"nova-cell0-cell-mapping-qxkzj\" (UID: \"94199ff4-bea2-4054-9752-7f9e5cd73d99\") " pod="openstack/nova-cell0-cell-mapping-qxkzj" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.450347 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94199ff4-bea2-4054-9752-7f9e5cd73d99-scripts\") pod \"nova-cell0-cell-mapping-qxkzj\" (UID: \"94199ff4-bea2-4054-9752-7f9e5cd73d99\") " pod="openstack/nova-cell0-cell-mapping-qxkzj" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.453039 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.454616 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-976zx\" (UniqueName: \"kubernetes.io/projected/94199ff4-bea2-4054-9752-7f9e5cd73d99-kube-api-access-976zx\") pod \"nova-cell0-cell-mapping-qxkzj\" (UID: \"94199ff4-bea2-4054-9752-7f9e5cd73d99\") " pod="openstack/nova-cell0-cell-mapping-qxkzj" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.454742 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.460489 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94199ff4-bea2-4054-9752-7f9e5cd73d99-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-qxkzj\" (UID: \"94199ff4-bea2-4054-9752-7f9e5cd73d99\") " pod="openstack/nova-cell0-cell-mapping-qxkzj" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.462817 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.467673 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"600b5343-a21b-4c61-94b4-54678000d8de","Type":"ContainerStarted","Data":"9674809e36b8880f24e5e524b5c8bd1353ae482612663892a3e294244f5de59f"} Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.468308 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.468527 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.518479 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qxkzj" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.528835 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f0f40ab-1168-4fea-84b1-b134eb2a1b3c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0f0f40ab-1168-4fea-84b1-b134eb2a1b3c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.528908 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f0f40ab-1168-4fea-84b1-b134eb2a1b3c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0f0f40ab-1168-4fea-84b1-b134eb2a1b3c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.528929 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/794826da-01ec-4ee2-99dd-e5ece92df833-config-data\") pod \"nova-scheduler-0\" (UID: \"794826da-01ec-4ee2-99dd-e5ece92df833\") " pod="openstack/nova-scheduler-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.529019 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpdgk\" (UniqueName: \"kubernetes.io/projected/0f0f40ab-1168-4fea-84b1-b134eb2a1b3c-kube-api-access-zpdgk\") pod \"nova-cell1-novncproxy-0\" (UID: \"0f0f40ab-1168-4fea-84b1-b134eb2a1b3c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.529176 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/794826da-01ec-4ee2-99dd-e5ece92df833-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"794826da-01ec-4ee2-99dd-e5ece92df833\") " pod="openstack/nova-scheduler-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.529206 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nptb\" (UniqueName: \"kubernetes.io/projected/794826da-01ec-4ee2-99dd-e5ece92df833-kube-api-access-7nptb\") pod \"nova-scheduler-0\" (UID: \"794826da-01ec-4ee2-99dd-e5ece92df833\") " pod="openstack/nova-scheduler-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.543734 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/794826da-01ec-4ee2-99dd-e5ece92df833-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"794826da-01ec-4ee2-99dd-e5ece92df833\") " pod="openstack/nova-scheduler-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.565052 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/794826da-01ec-4ee2-99dd-e5ece92df833-config-data\") pod \"nova-scheduler-0\" (UID: \"794826da-01ec-4ee2-99dd-e5ece92df833\") " pod="openstack/nova-scheduler-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.566642 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nptb\" (UniqueName: \"kubernetes.io/projected/794826da-01ec-4ee2-99dd-e5ece92df833-kube-api-access-7nptb\") pod \"nova-scheduler-0\" (UID: \"794826da-01ec-4ee2-99dd-e5ece92df833\") " pod="openstack/nova-scheduler-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.608919 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.611024 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.613737 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.634193 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f0f40ab-1168-4fea-84b1-b134eb2a1b3c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0f0f40ab-1168-4fea-84b1-b134eb2a1b3c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.634248 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f0f40ab-1168-4fea-84b1-b134eb2a1b3c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0f0f40ab-1168-4fea-84b1-b134eb2a1b3c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.634323 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpdgk\" (UniqueName: \"kubernetes.io/projected/0f0f40ab-1168-4fea-84b1-b134eb2a1b3c-kube-api-access-zpdgk\") pod \"nova-cell1-novncproxy-0\" (UID: \"0f0f40ab-1168-4fea-84b1-b134eb2a1b3c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.637632 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.639011 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f0f40ab-1168-4fea-84b1-b134eb2a1b3c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"0f0f40ab-1168-4fea-84b1-b134eb2a1b3c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.639446 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.639485 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.173017979 podStartE2EDuration="5.639476514s" podCreationTimestamp="2025-12-27 08:06:31 +0000 UTC" firstStartedPulling="2025-12-27 08:06:32.335714787 +0000 UTC m=+1453.156155391" lastFinishedPulling="2025-12-27 08:06:35.802173342 +0000 UTC m=+1456.622613926" observedRunningTime="2025-12-27 08:06:36.560763933 +0000 UTC m=+1457.381204527" watchObservedRunningTime="2025-12-27 08:06:36.639476514 +0000 UTC m=+1457.459917108" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.640859 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.643412 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f0f40ab-1168-4fea-84b1-b134eb2a1b3c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"0f0f40ab-1168-4fea-84b1-b134eb2a1b3c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.669833 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpdgk\" (UniqueName: \"kubernetes.io/projected/0f0f40ab-1168-4fea-84b1-b134eb2a1b3c-kube-api-access-zpdgk\") pod \"nova-cell1-novncproxy-0\" (UID: \"0f0f40ab-1168-4fea-84b1-b134eb2a1b3c\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.702186 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.711843 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.736431 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-942r2\" (UniqueName: \"kubernetes.io/projected/379fe74c-b561-44e6-8f06-2e0818922a09-kube-api-access-942r2\") pod \"nova-api-0\" (UID: \"379fe74c-b561-44e6-8f06-2e0818922a09\") " pod="openstack/nova-api-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.736476 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/379fe74c-b561-44e6-8f06-2e0818922a09-logs\") pod \"nova-api-0\" (UID: \"379fe74c-b561-44e6-8f06-2e0818922a09\") " pod="openstack/nova-api-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.736571 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/379fe74c-b561-44e6-8f06-2e0818922a09-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"379fe74c-b561-44e6-8f06-2e0818922a09\") " pod="openstack/nova-api-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.736614 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c696b92-7128-4465-875f-d2ef4298e798-logs\") pod \"nova-metadata-0\" (UID: \"4c696b92-7128-4465-875f-d2ef4298e798\") " pod="openstack/nova-metadata-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.736645 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c696b92-7128-4465-875f-d2ef4298e798-config-data\") pod \"nova-metadata-0\" (UID: \"4c696b92-7128-4465-875f-d2ef4298e798\") " pod="openstack/nova-metadata-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.736668 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thlkl\" (UniqueName: \"kubernetes.io/projected/4c696b92-7128-4465-875f-d2ef4298e798-kube-api-access-thlkl\") pod \"nova-metadata-0\" (UID: \"4c696b92-7128-4465-875f-d2ef4298e798\") " pod="openstack/nova-metadata-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.736717 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/379fe74c-b561-44e6-8f06-2e0818922a09-config-data\") pod \"nova-api-0\" (UID: \"379fe74c-b561-44e6-8f06-2e0818922a09\") " pod="openstack/nova-api-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.736765 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c696b92-7128-4465-875f-d2ef4298e798-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4c696b92-7128-4465-875f-d2ef4298e798\") " pod="openstack/nova-metadata-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.761044 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.801912 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-446hx"] Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.826914 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.853821 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-942r2\" (UniqueName: \"kubernetes.io/projected/379fe74c-b561-44e6-8f06-2e0818922a09-kube-api-access-942r2\") pod \"nova-api-0\" (UID: \"379fe74c-b561-44e6-8f06-2e0818922a09\") " pod="openstack/nova-api-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.853911 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-ovsdbserver-nb\") pod \"dnsmasq-dns-568d7fd7cf-446hx\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.853940 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wtck\" (UniqueName: \"kubernetes.io/projected/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-kube-api-access-7wtck\") pod \"dnsmasq-dns-568d7fd7cf-446hx\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.853973 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/379fe74c-b561-44e6-8f06-2e0818922a09-logs\") pod \"nova-api-0\" (UID: \"379fe74c-b561-44e6-8f06-2e0818922a09\") " pod="openstack/nova-api-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.854130 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-dns-swift-storage-0\") pod \"dnsmasq-dns-568d7fd7cf-446hx\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.870930 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/379fe74c-b561-44e6-8f06-2e0818922a09-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"379fe74c-b561-44e6-8f06-2e0818922a09\") " pod="openstack/nova-api-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.871076 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c696b92-7128-4465-875f-d2ef4298e798-logs\") pod \"nova-metadata-0\" (UID: \"4c696b92-7128-4465-875f-d2ef4298e798\") " pod="openstack/nova-metadata-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.871174 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c696b92-7128-4465-875f-d2ef4298e798-config-data\") pod \"nova-metadata-0\" (UID: \"4c696b92-7128-4465-875f-d2ef4298e798\") " pod="openstack/nova-metadata-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.871222 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thlkl\" (UniqueName: \"kubernetes.io/projected/4c696b92-7128-4465-875f-d2ef4298e798-kube-api-access-thlkl\") pod \"nova-metadata-0\" (UID: \"4c696b92-7128-4465-875f-d2ef4298e798\") " pod="openstack/nova-metadata-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.871313 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-dns-svc\") pod \"dnsmasq-dns-568d7fd7cf-446hx\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.871347 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-config\") pod \"dnsmasq-dns-568d7fd7cf-446hx\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.871436 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/379fe74c-b561-44e6-8f06-2e0818922a09-config-data\") pod \"nova-api-0\" (UID: \"379fe74c-b561-44e6-8f06-2e0818922a09\") " pod="openstack/nova-api-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.871533 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c696b92-7128-4465-875f-d2ef4298e798-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4c696b92-7128-4465-875f-d2ef4298e798\") " pod="openstack/nova-metadata-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.871581 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-ovsdbserver-sb\") pod \"dnsmasq-dns-568d7fd7cf-446hx\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.872149 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/379fe74c-b561-44e6-8f06-2e0818922a09-logs\") pod \"nova-api-0\" (UID: \"379fe74c-b561-44e6-8f06-2e0818922a09\") " pod="openstack/nova-api-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.878887 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c696b92-7128-4465-875f-d2ef4298e798-logs\") pod \"nova-metadata-0\" (UID: \"4c696b92-7128-4465-875f-d2ef4298e798\") " pod="openstack/nova-metadata-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.885591 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-942r2\" (UniqueName: \"kubernetes.io/projected/379fe74c-b561-44e6-8f06-2e0818922a09-kube-api-access-942r2\") pod \"nova-api-0\" (UID: \"379fe74c-b561-44e6-8f06-2e0818922a09\") " pod="openstack/nova-api-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.889552 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-446hx"] Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.894407 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c696b92-7128-4465-875f-d2ef4298e798-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4c696b92-7128-4465-875f-d2ef4298e798\") " pod="openstack/nova-metadata-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.913026 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c696b92-7128-4465-875f-d2ef4298e798-config-data\") pod \"nova-metadata-0\" (UID: \"4c696b92-7128-4465-875f-d2ef4298e798\") " pod="openstack/nova-metadata-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.917727 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/379fe74c-b561-44e6-8f06-2e0818922a09-config-data\") pod \"nova-api-0\" (UID: \"379fe74c-b561-44e6-8f06-2e0818922a09\") " pod="openstack/nova-api-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.925373 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thlkl\" (UniqueName: \"kubernetes.io/projected/4c696b92-7128-4465-875f-d2ef4298e798-kube-api-access-thlkl\") pod \"nova-metadata-0\" (UID: \"4c696b92-7128-4465-875f-d2ef4298e798\") " pod="openstack/nova-metadata-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.929948 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/379fe74c-b561-44e6-8f06-2e0818922a09-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"379fe74c-b561-44e6-8f06-2e0818922a09\") " pod="openstack/nova-api-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.971686 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.973255 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-dns-svc\") pod \"dnsmasq-dns-568d7fd7cf-446hx\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.973299 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-config\") pod \"dnsmasq-dns-568d7fd7cf-446hx\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.973365 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-ovsdbserver-sb\") pod \"dnsmasq-dns-568d7fd7cf-446hx\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.973411 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-ovsdbserver-nb\") pod \"dnsmasq-dns-568d7fd7cf-446hx\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.973430 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wtck\" (UniqueName: \"kubernetes.io/projected/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-kube-api-access-7wtck\") pod \"dnsmasq-dns-568d7fd7cf-446hx\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.973466 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-dns-swift-storage-0\") pod \"dnsmasq-dns-568d7fd7cf-446hx\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.976505 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-ovsdbserver-sb\") pod \"dnsmasq-dns-568d7fd7cf-446hx\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.978690 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-config\") pod \"dnsmasq-dns-568d7fd7cf-446hx\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.977781 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-dns-svc\") pod \"dnsmasq-dns-568d7fd7cf-446hx\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.979460 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-ovsdbserver-nb\") pod \"dnsmasq-dns-568d7fd7cf-446hx\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.980822 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-dns-swift-storage-0\") pod \"dnsmasq-dns-568d7fd7cf-446hx\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:36 crc kubenswrapper[4934]: I1227 08:06:36.986930 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.010923 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wtck\" (UniqueName: \"kubernetes.io/projected/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-kube-api-access-7wtck\") pod \"dnsmasq-dns-568d7fd7cf-446hx\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.012547 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.151295 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.319832 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-4vl7v" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.384150 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l76wt\" (UniqueName: \"kubernetes.io/projected/38941274-cf93-4985-9e25-f9aad3dcb394-kube-api-access-l76wt\") pod \"38941274-cf93-4985-9e25-f9aad3dcb394\" (UID: \"38941274-cf93-4985-9e25-f9aad3dcb394\") " Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.393033 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38941274-cf93-4985-9e25-f9aad3dcb394-kube-api-access-l76wt" (OuterVolumeSpecName: "kube-api-access-l76wt") pod "38941274-cf93-4985-9e25-f9aad3dcb394" (UID: "38941274-cf93-4985-9e25-f9aad3dcb394"). InnerVolumeSpecName "kube-api-access-l76wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.486585 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l76wt\" (UniqueName: \"kubernetes.io/projected/38941274-cf93-4985-9e25-f9aad3dcb394-kube-api-access-l76wt\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.498652 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-4vl7v" event={"ID":"38941274-cf93-4985-9e25-f9aad3dcb394","Type":"ContainerDied","Data":"c538d5506fb7899223a0e0a79e0333b18d418a2aa998011d8163c109cf8eecb3"} Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.498722 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c538d5506fb7899223a0e0a79e0333b18d418a2aa998011d8163c109cf8eecb3" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.498860 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-4vl7v" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.521418 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.550060 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-qxkzj"] Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.683441 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6kzb4"] Dec 27 08:06:37 crc kubenswrapper[4934]: E1227 08:06:37.684106 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38941274-cf93-4985-9e25-f9aad3dcb394" containerName="mariadb-database-create" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.684131 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="38941274-cf93-4985-9e25-f9aad3dcb394" containerName="mariadb-database-create" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.684482 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="38941274-cf93-4985-9e25-f9aad3dcb394" containerName="mariadb-database-create" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.685573 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6kzb4" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.687664 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.688421 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.696481 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6kzb4"] Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.793182 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct489\" (UniqueName: \"kubernetes.io/projected/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-kube-api-access-ct489\") pod \"nova-cell1-conductor-db-sync-6kzb4\" (UID: \"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e\") " pod="openstack/nova-cell1-conductor-db-sync-6kzb4" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.793226 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-6kzb4\" (UID: \"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e\") " pod="openstack/nova-cell1-conductor-db-sync-6kzb4" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.793314 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-scripts\") pod \"nova-cell1-conductor-db-sync-6kzb4\" (UID: \"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e\") " pod="openstack/nova-cell1-conductor-db-sync-6kzb4" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.793345 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-config-data\") pod \"nova-cell1-conductor-db-sync-6kzb4\" (UID: \"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e\") " pod="openstack/nova-cell1-conductor-db-sync-6kzb4" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.820936 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.899811 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-6kzb4\" (UID: \"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e\") " pod="openstack/nova-cell1-conductor-db-sync-6kzb4" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.899865 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct489\" (UniqueName: \"kubernetes.io/projected/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-kube-api-access-ct489\") pod \"nova-cell1-conductor-db-sync-6kzb4\" (UID: \"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e\") " pod="openstack/nova-cell1-conductor-db-sync-6kzb4" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.900007 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-scripts\") pod \"nova-cell1-conductor-db-sync-6kzb4\" (UID: \"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e\") " pod="openstack/nova-cell1-conductor-db-sync-6kzb4" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.900053 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-config-data\") pod \"nova-cell1-conductor-db-sync-6kzb4\" (UID: \"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e\") " pod="openstack/nova-cell1-conductor-db-sync-6kzb4" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.904966 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-scripts\") pod \"nova-cell1-conductor-db-sync-6kzb4\" (UID: \"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e\") " pod="openstack/nova-cell1-conductor-db-sync-6kzb4" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.905503 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-6kzb4\" (UID: \"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e\") " pod="openstack/nova-cell1-conductor-db-sync-6kzb4" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.909189 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-config-data\") pod \"nova-cell1-conductor-db-sync-6kzb4\" (UID: \"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e\") " pod="openstack/nova-cell1-conductor-db-sync-6kzb4" Dec 27 08:06:37 crc kubenswrapper[4934]: I1227 08:06:37.923729 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct489\" (UniqueName: \"kubernetes.io/projected/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-kube-api-access-ct489\") pod \"nova-cell1-conductor-db-sync-6kzb4\" (UID: \"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e\") " pod="openstack/nova-cell1-conductor-db-sync-6kzb4" Dec 27 08:06:38 crc kubenswrapper[4934]: I1227 08:06:38.030431 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 27 08:06:38 crc kubenswrapper[4934]: W1227 08:06:38.032208 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod379fe74c_b561_44e6_8f06_2e0818922a09.slice/crio-05df7fc936b43cab07fa0a1bdbb2fda7d3b7213529104e34be128889c453e9e9 WatchSource:0}: Error finding container 05df7fc936b43cab07fa0a1bdbb2fda7d3b7213529104e34be128889c453e9e9: Status 404 returned error can't find the container with id 05df7fc936b43cab07fa0a1bdbb2fda7d3b7213529104e34be128889c453e9e9 Dec 27 08:06:38 crc kubenswrapper[4934]: I1227 08:06:38.041938 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6kzb4" Dec 27 08:06:38 crc kubenswrapper[4934]: I1227 08:06:38.106522 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-446hx"] Dec 27 08:06:38 crc kubenswrapper[4934]: I1227 08:06:38.125661 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 27 08:06:38 crc kubenswrapper[4934]: I1227 08:06:38.530388 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" event={"ID":"d1ddee33-e9fa-4651-b7d6-d39fa59b542c","Type":"ContainerStarted","Data":"bf3d59f1cfb3404d87c6c0548624865cbc071697bdacbdcad9ee17fbfc0486aa"} Dec 27 08:06:38 crc kubenswrapper[4934]: I1227 08:06:38.542541 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qxkzj" event={"ID":"94199ff4-bea2-4054-9752-7f9e5cd73d99","Type":"ContainerStarted","Data":"b2b5df90fe5da94b89d952defbf811bc55a7d5d080faf0300c01ddbb0551a5d1"} Dec 27 08:06:38 crc kubenswrapper[4934]: I1227 08:06:38.542585 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qxkzj" event={"ID":"94199ff4-bea2-4054-9752-7f9e5cd73d99","Type":"ContainerStarted","Data":"62afcb49a7b971c15c356e15a05d88355ab7581c5b339868724e2faacd2bc886"} Dec 27 08:06:38 crc kubenswrapper[4934]: I1227 08:06:38.553442 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0f0f40ab-1168-4fea-84b1-b134eb2a1b3c","Type":"ContainerStarted","Data":"d3712aea3b2913aa199387cdf9f87000c6483d558814a9708bc3320132fb6272"} Dec 27 08:06:38 crc kubenswrapper[4934]: I1227 08:06:38.555045 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"794826da-01ec-4ee2-99dd-e5ece92df833","Type":"ContainerStarted","Data":"f152f356505fab18d34805b6c516f652628fa69640f40a31c4b9aab41b9159f2"} Dec 27 08:06:38 crc kubenswrapper[4934]: I1227 08:06:38.557628 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c696b92-7128-4465-875f-d2ef4298e798","Type":"ContainerStarted","Data":"1a78506bb3f041e1daf078de7f37bad24b4eca3082f9942981c516c2990b591f"} Dec 27 08:06:38 crc kubenswrapper[4934]: I1227 08:06:38.572688 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"379fe74c-b561-44e6-8f06-2e0818922a09","Type":"ContainerStarted","Data":"05df7fc936b43cab07fa0a1bdbb2fda7d3b7213529104e34be128889c453e9e9"} Dec 27 08:06:38 crc kubenswrapper[4934]: I1227 08:06:38.584518 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-qxkzj" podStartSLOduration=2.584495512 podStartE2EDuration="2.584495512s" podCreationTimestamp="2025-12-27 08:06:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:06:38.572896576 +0000 UTC m=+1459.393337170" watchObservedRunningTime="2025-12-27 08:06:38.584495512 +0000 UTC m=+1459.404936096" Dec 27 08:06:38 crc kubenswrapper[4934]: I1227 08:06:38.697613 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6kzb4"] Dec 27 08:06:38 crc kubenswrapper[4934]: W1227 08:06:38.708648 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52a3f24b_77b1_467e_b9ce_03e2ae5fe53e.slice/crio-9e0252c49143b7df4742f58d1f774094c51836957ffc5d715ccc0594ea80f313 WatchSource:0}: Error finding container 9e0252c49143b7df4742f58d1f774094c51836957ffc5d715ccc0594ea80f313: Status 404 returned error can't find the container with id 9e0252c49143b7df4742f58d1f774094c51836957ffc5d715ccc0594ea80f313 Dec 27 08:06:39 crc kubenswrapper[4934]: I1227 08:06:39.791071 4934 generic.go:334] "Generic (PLEG): container finished" podID="d1ddee33-e9fa-4651-b7d6-d39fa59b542c" containerID="e4d6d1872e20de0be62ae2eb22ebed27b2ce0b757517e9b1c9d6688180b9bcbd" exitCode=0 Dec 27 08:06:39 crc kubenswrapper[4934]: I1227 08:06:39.816362 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6kzb4" event={"ID":"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e","Type":"ContainerStarted","Data":"201c806d5834ec84e492faab44bdf7372b91b98e98442e86462b31551fb92d58"} Dec 27 08:06:39 crc kubenswrapper[4934]: I1227 08:06:39.816452 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6kzb4" event={"ID":"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e","Type":"ContainerStarted","Data":"9e0252c49143b7df4742f58d1f774094c51836957ffc5d715ccc0594ea80f313"} Dec 27 08:06:39 crc kubenswrapper[4934]: I1227 08:06:39.816469 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" event={"ID":"d1ddee33-e9fa-4651-b7d6-d39fa59b542c","Type":"ContainerDied","Data":"e4d6d1872e20de0be62ae2eb22ebed27b2ce0b757517e9b1c9d6688180b9bcbd"} Dec 27 08:06:39 crc kubenswrapper[4934]: I1227 08:06:39.942388 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-6kzb4" podStartSLOduration=2.942369486 podStartE2EDuration="2.942369486s" podCreationTimestamp="2025-12-27 08:06:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:06:39.921888551 +0000 UTC m=+1460.742329145" watchObservedRunningTime="2025-12-27 08:06:39.942369486 +0000 UTC m=+1460.762810080" Dec 27 08:06:40 crc kubenswrapper[4934]: I1227 08:06:40.150010 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 27 08:06:40 crc kubenswrapper[4934]: I1227 08:06:40.164586 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 27 08:06:42 crc kubenswrapper[4934]: I1227 08:06:42.825512 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"794826da-01ec-4ee2-99dd-e5ece92df833","Type":"ContainerStarted","Data":"236fa6ccd4c0ae3a62409338ff24ba79be6a75b377d52ba6570f2da9651436b3"} Dec 27 08:06:42 crc kubenswrapper[4934]: I1227 08:06:42.835902 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c696b92-7128-4465-875f-d2ef4298e798","Type":"ContainerStarted","Data":"bc9b8337126f66220fa9bcfe2a59c8bf4423fcc2292e6bbc9bf9ca725a983c0b"} Dec 27 08:06:42 crc kubenswrapper[4934]: I1227 08:06:42.835962 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c696b92-7128-4465-875f-d2ef4298e798","Type":"ContainerStarted","Data":"0e363c5e450ac1ceba74fb25171c16da42c41caad9e6478a642ac132a0d9e0af"} Dec 27 08:06:42 crc kubenswrapper[4934]: I1227 08:06:42.836120 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4c696b92-7128-4465-875f-d2ef4298e798" containerName="nova-metadata-log" containerID="cri-o://0e363c5e450ac1ceba74fb25171c16da42c41caad9e6478a642ac132a0d9e0af" gracePeriod=30 Dec 27 08:06:42 crc kubenswrapper[4934]: I1227 08:06:42.836406 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4c696b92-7128-4465-875f-d2ef4298e798" containerName="nova-metadata-metadata" containerID="cri-o://bc9b8337126f66220fa9bcfe2a59c8bf4423fcc2292e6bbc9bf9ca725a983c0b" gracePeriod=30 Dec 27 08:06:42 crc kubenswrapper[4934]: I1227 08:06:42.838896 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"379fe74c-b561-44e6-8f06-2e0818922a09","Type":"ContainerStarted","Data":"601d6356bc1737811808ec2361391369064bbccc90c3a004c42c79af9f7c5c98"} Dec 27 08:06:42 crc kubenswrapper[4934]: I1227 08:06:42.838924 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"379fe74c-b561-44e6-8f06-2e0818922a09","Type":"ContainerStarted","Data":"b91ec4ad0a607ccb9f99f9f8a582817bcb77df25419ae2f0833daea69eb14b87"} Dec 27 08:06:42 crc kubenswrapper[4934]: I1227 08:06:42.843661 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" event={"ID":"d1ddee33-e9fa-4651-b7d6-d39fa59b542c","Type":"ContainerStarted","Data":"e36a87cfc635581d28738c58261e1b781bc8daaa5d652e75c956ee44f66f77f8"} Dec 27 08:06:42 crc kubenswrapper[4934]: I1227 08:06:42.843818 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:42 crc kubenswrapper[4934]: I1227 08:06:42.845810 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0f0f40ab-1168-4fea-84b1-b134eb2a1b3c","Type":"ContainerStarted","Data":"bc6848ec46001daa30195a2f745ad7629c7bf68dd56b59d8b02fda09322ce597"} Dec 27 08:06:42 crc kubenswrapper[4934]: I1227 08:06:42.845946 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="0f0f40ab-1168-4fea-84b1-b134eb2a1b3c" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://bc6848ec46001daa30195a2f745ad7629c7bf68dd56b59d8b02fda09322ce597" gracePeriod=30 Dec 27 08:06:42 crc kubenswrapper[4934]: I1227 08:06:42.862122 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.366877837 podStartE2EDuration="6.861884078s" podCreationTimestamp="2025-12-27 08:06:36 +0000 UTC" firstStartedPulling="2025-12-27 08:06:37.553448224 +0000 UTC m=+1458.373888818" lastFinishedPulling="2025-12-27 08:06:42.048454465 +0000 UTC m=+1462.868895059" observedRunningTime="2025-12-27 08:06:42.854433034 +0000 UTC m=+1463.674873628" watchObservedRunningTime="2025-12-27 08:06:42.861884078 +0000 UTC m=+1463.682324672" Dec 27 08:06:42 crc kubenswrapper[4934]: I1227 08:06:42.951961 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.720103375 podStartE2EDuration="6.951926868s" podCreationTimestamp="2025-12-27 08:06:36 +0000 UTC" firstStartedPulling="2025-12-27 08:06:37.829393847 +0000 UTC m=+1458.649834441" lastFinishedPulling="2025-12-27 08:06:42.06121734 +0000 UTC m=+1462.881657934" observedRunningTime="2025-12-27 08:06:42.88305482 +0000 UTC m=+1463.703495414" watchObservedRunningTime="2025-12-27 08:06:42.951926868 +0000 UTC m=+1463.772367462" Dec 27 08:06:42 crc kubenswrapper[4934]: I1227 08:06:42.964974 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.068192665 podStartE2EDuration="6.964956629s" podCreationTimestamp="2025-12-27 08:06:36 +0000 UTC" firstStartedPulling="2025-12-27 08:06:38.16176387 +0000 UTC m=+1458.982204464" lastFinishedPulling="2025-12-27 08:06:42.058527824 +0000 UTC m=+1462.878968428" observedRunningTime="2025-12-27 08:06:42.913565082 +0000 UTC m=+1463.734005666" watchObservedRunningTime="2025-12-27 08:06:42.964956629 +0000 UTC m=+1463.785397223" Dec 27 08:06:42 crc kubenswrapper[4934]: I1227 08:06:42.975688 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.948642689 podStartE2EDuration="6.975673923s" podCreationTimestamp="2025-12-27 08:06:36 +0000 UTC" firstStartedPulling="2025-12-27 08:06:38.034930824 +0000 UTC m=+1458.855371418" lastFinishedPulling="2025-12-27 08:06:42.061962038 +0000 UTC m=+1462.882402652" observedRunningTime="2025-12-27 08:06:42.941101201 +0000 UTC m=+1463.761541795" watchObservedRunningTime="2025-12-27 08:06:42.975673923 +0000 UTC m=+1463.796114517" Dec 27 08:06:42 crc kubenswrapper[4934]: I1227 08:06:42.985778 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" podStartSLOduration=6.985754482 podStartE2EDuration="6.985754482s" podCreationTimestamp="2025-12-27 08:06:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:06:42.960210542 +0000 UTC m=+1463.780651136" watchObservedRunningTime="2025-12-27 08:06:42.985754482 +0000 UTC m=+1463.806195076" Dec 27 08:06:43 crc kubenswrapper[4934]: I1227 08:06:43.751170 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-2566-account-create-rtm45"] Dec 27 08:06:43 crc kubenswrapper[4934]: I1227 08:06:43.753318 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-2566-account-create-rtm45" Dec 27 08:06:43 crc kubenswrapper[4934]: I1227 08:06:43.755659 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Dec 27 08:06:43 crc kubenswrapper[4934]: I1227 08:06:43.765840 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-2566-account-create-rtm45"] Dec 27 08:06:43 crc kubenswrapper[4934]: I1227 08:06:43.858635 4934 generic.go:334] "Generic (PLEG): container finished" podID="4c696b92-7128-4465-875f-d2ef4298e798" containerID="0e363c5e450ac1ceba74fb25171c16da42c41caad9e6478a642ac132a0d9e0af" exitCode=143 Dec 27 08:06:43 crc kubenswrapper[4934]: I1227 08:06:43.859575 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c696b92-7128-4465-875f-d2ef4298e798","Type":"ContainerDied","Data":"0e363c5e450ac1ceba74fb25171c16da42c41caad9e6478a642ac132a0d9e0af"} Dec 27 08:06:43 crc kubenswrapper[4934]: I1227 08:06:43.875985 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlldf\" (UniqueName: \"kubernetes.io/projected/72321394-c1c2-4254-9e2a-10617a3ba8c4-kube-api-access-rlldf\") pod \"aodh-2566-account-create-rtm45\" (UID: \"72321394-c1c2-4254-9e2a-10617a3ba8c4\") " pod="openstack/aodh-2566-account-create-rtm45" Dec 27 08:06:43 crc kubenswrapper[4934]: I1227 08:06:43.979318 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlldf\" (UniqueName: \"kubernetes.io/projected/72321394-c1c2-4254-9e2a-10617a3ba8c4-kube-api-access-rlldf\") pod \"aodh-2566-account-create-rtm45\" (UID: \"72321394-c1c2-4254-9e2a-10617a3ba8c4\") " pod="openstack/aodh-2566-account-create-rtm45" Dec 27 08:06:44 crc kubenswrapper[4934]: I1227 08:06:44.000592 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlldf\" (UniqueName: \"kubernetes.io/projected/72321394-c1c2-4254-9e2a-10617a3ba8c4-kube-api-access-rlldf\") pod \"aodh-2566-account-create-rtm45\" (UID: \"72321394-c1c2-4254-9e2a-10617a3ba8c4\") " pod="openstack/aodh-2566-account-create-rtm45" Dec 27 08:06:44 crc kubenswrapper[4934]: I1227 08:06:44.076614 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-2566-account-create-rtm45" Dec 27 08:06:44 crc kubenswrapper[4934]: I1227 08:06:44.626237 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-2566-account-create-rtm45"] Dec 27 08:06:44 crc kubenswrapper[4934]: I1227 08:06:44.870698 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-2566-account-create-rtm45" event={"ID":"72321394-c1c2-4254-9e2a-10617a3ba8c4","Type":"ContainerStarted","Data":"a3ba409c30638f31737d02cdbbb7f3b230286a84cb70201714fe63efda870e2c"} Dec 27 08:06:44 crc kubenswrapper[4934]: I1227 08:06:44.871023 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-2566-account-create-rtm45" event={"ID":"72321394-c1c2-4254-9e2a-10617a3ba8c4","Type":"ContainerStarted","Data":"544f0e50c3a9d498105233c322d492907b3e1b8765baea78ec8ca43052330b72"} Dec 27 08:06:44 crc kubenswrapper[4934]: I1227 08:06:44.889858 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-2566-account-create-rtm45" podStartSLOduration=1.889839751 podStartE2EDuration="1.889839751s" podCreationTimestamp="2025-12-27 08:06:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:06:44.883559756 +0000 UTC m=+1465.704000360" watchObservedRunningTime="2025-12-27 08:06:44.889839751 +0000 UTC m=+1465.710280335" Dec 27 08:06:45 crc kubenswrapper[4934]: I1227 08:06:45.904902 4934 generic.go:334] "Generic (PLEG): container finished" podID="72321394-c1c2-4254-9e2a-10617a3ba8c4" containerID="a3ba409c30638f31737d02cdbbb7f3b230286a84cb70201714fe63efda870e2c" exitCode=0 Dec 27 08:06:45 crc kubenswrapper[4934]: I1227 08:06:45.904988 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-2566-account-create-rtm45" event={"ID":"72321394-c1c2-4254-9e2a-10617a3ba8c4","Type":"ContainerDied","Data":"a3ba409c30638f31737d02cdbbb7f3b230286a84cb70201714fe63efda870e2c"} Dec 27 08:06:46 crc kubenswrapper[4934]: I1227 08:06:46.703408 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 27 08:06:46 crc kubenswrapper[4934]: I1227 08:06:46.703957 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 27 08:06:46 crc kubenswrapper[4934]: I1227 08:06:46.764731 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 27 08:06:46 crc kubenswrapper[4934]: I1227 08:06:46.922123 4934 generic.go:334] "Generic (PLEG): container finished" podID="52a3f24b-77b1-467e-b9ce-03e2ae5fe53e" containerID="201c806d5834ec84e492faab44bdf7372b91b98e98442e86462b31551fb92d58" exitCode=0 Dec 27 08:06:46 crc kubenswrapper[4934]: I1227 08:06:46.922185 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6kzb4" event={"ID":"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e","Type":"ContainerDied","Data":"201c806d5834ec84e492faab44bdf7372b91b98e98442e86462b31551fb92d58"} Dec 27 08:06:46 crc kubenswrapper[4934]: I1227 08:06:46.924192 4934 generic.go:334] "Generic (PLEG): container finished" podID="94199ff4-bea2-4054-9752-7f9e5cd73d99" containerID="b2b5df90fe5da94b89d952defbf811bc55a7d5d080faf0300c01ddbb0551a5d1" exitCode=0 Dec 27 08:06:46 crc kubenswrapper[4934]: I1227 08:06:46.924265 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qxkzj" event={"ID":"94199ff4-bea2-4054-9752-7f9e5cd73d99","Type":"ContainerDied","Data":"b2b5df90fe5da94b89d952defbf811bc55a7d5d080faf0300c01ddbb0551a5d1"} Dec 27 08:06:46 crc kubenswrapper[4934]: I1227 08:06:46.972572 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:06:46 crc kubenswrapper[4934]: I1227 08:06:46.973759 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 27 08:06:46 crc kubenswrapper[4934]: I1227 08:06:46.989469 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 27 08:06:46 crc kubenswrapper[4934]: I1227 08:06:46.989933 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.014398 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.014483 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.154296 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.238558 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-l22fd"] Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.238854 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" podUID="ee5d7a6d-2acf-4cfa-b028-de1c3a20c546" containerName="dnsmasq-dns" containerID="cri-o://91212881dafecfe654aa3a58c9f80410ec17d93fb18b80c38eb052ad42b89c8a" gracePeriod=10 Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.434894 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-2566-account-create-rtm45" Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.573889 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlldf\" (UniqueName: \"kubernetes.io/projected/72321394-c1c2-4254-9e2a-10617a3ba8c4-kube-api-access-rlldf\") pod \"72321394-c1c2-4254-9e2a-10617a3ba8c4\" (UID: \"72321394-c1c2-4254-9e2a-10617a3ba8c4\") " Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.585420 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72321394-c1c2-4254-9e2a-10617a3ba8c4-kube-api-access-rlldf" (OuterVolumeSpecName: "kube-api-access-rlldf") pod "72321394-c1c2-4254-9e2a-10617a3ba8c4" (UID: "72321394-c1c2-4254-9e2a-10617a3ba8c4"). InnerVolumeSpecName "kube-api-access-rlldf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.680754 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlldf\" (UniqueName: \"kubernetes.io/projected/72321394-c1c2-4254-9e2a-10617a3ba8c4-kube-api-access-rlldf\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.877003 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.954561 4934 generic.go:334] "Generic (PLEG): container finished" podID="ee5d7a6d-2acf-4cfa-b028-de1c3a20c546" containerID="91212881dafecfe654aa3a58c9f80410ec17d93fb18b80c38eb052ad42b89c8a" exitCode=0 Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.954617 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" event={"ID":"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546","Type":"ContainerDied","Data":"91212881dafecfe654aa3a58c9f80410ec17d93fb18b80c38eb052ad42b89c8a"} Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.954644 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" event={"ID":"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546","Type":"ContainerDied","Data":"ecaa5af55972576661cc763665718fa1c11cce6d34ed7ab546a465a3f6650dbc"} Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.954659 4934 scope.go:117] "RemoveContainer" containerID="91212881dafecfe654aa3a58c9f80410ec17d93fb18b80c38eb052ad42b89c8a" Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.954789 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.960771 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-2566-account-create-rtm45" Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.960837 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-2566-account-create-rtm45" event={"ID":"72321394-c1c2-4254-9e2a-10617a3ba8c4","Type":"ContainerDied","Data":"544f0e50c3a9d498105233c322d492907b3e1b8765baea78ec8ca43052330b72"} Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.960864 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="544f0e50c3a9d498105233c322d492907b3e1b8765baea78ec8ca43052330b72" Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.987652 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpk9k\" (UniqueName: \"kubernetes.io/projected/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-kube-api-access-mpk9k\") pod \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.987980 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-dns-svc\") pod \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.988010 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-config\") pod \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.988138 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-ovsdbserver-sb\") pod \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.988183 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-dns-swift-storage-0\") pod \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.988312 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-ovsdbserver-nb\") pod \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\" (UID: \"ee5d7a6d-2acf-4cfa-b028-de1c3a20c546\") " Dec 27 08:06:47 crc kubenswrapper[4934]: I1227 08:06:47.995318 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-kube-api-access-mpk9k" (OuterVolumeSpecName: "kube-api-access-mpk9k") pod "ee5d7a6d-2acf-4cfa-b028-de1c3a20c546" (UID: "ee5d7a6d-2acf-4cfa-b028-de1c3a20c546"). InnerVolumeSpecName "kube-api-access-mpk9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.015946 4934 scope.go:117] "RemoveContainer" containerID="c4c45b7d7f63de8bfeca8e55a5b3a1096dd3359a49a71273e20ab35c776b7dc5" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.051975 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ee5d7a6d-2acf-4cfa-b028-de1c3a20c546" (UID: "ee5d7a6d-2acf-4cfa-b028-de1c3a20c546"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.054636 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ee5d7a6d-2acf-4cfa-b028-de1c3a20c546" (UID: "ee5d7a6d-2acf-4cfa-b028-de1c3a20c546"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.058188 4934 scope.go:117] "RemoveContainer" containerID="91212881dafecfe654aa3a58c9f80410ec17d93fb18b80c38eb052ad42b89c8a" Dec 27 08:06:48 crc kubenswrapper[4934]: E1227 08:06:48.058684 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91212881dafecfe654aa3a58c9f80410ec17d93fb18b80c38eb052ad42b89c8a\": container with ID starting with 91212881dafecfe654aa3a58c9f80410ec17d93fb18b80c38eb052ad42b89c8a not found: ID does not exist" containerID="91212881dafecfe654aa3a58c9f80410ec17d93fb18b80c38eb052ad42b89c8a" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.058731 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91212881dafecfe654aa3a58c9f80410ec17d93fb18b80c38eb052ad42b89c8a"} err="failed to get container status \"91212881dafecfe654aa3a58c9f80410ec17d93fb18b80c38eb052ad42b89c8a\": rpc error: code = NotFound desc = could not find container \"91212881dafecfe654aa3a58c9f80410ec17d93fb18b80c38eb052ad42b89c8a\": container with ID starting with 91212881dafecfe654aa3a58c9f80410ec17d93fb18b80c38eb052ad42b89c8a not found: ID does not exist" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.058764 4934 scope.go:117] "RemoveContainer" containerID="c4c45b7d7f63de8bfeca8e55a5b3a1096dd3359a49a71273e20ab35c776b7dc5" Dec 27 08:06:48 crc kubenswrapper[4934]: E1227 08:06:48.059367 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4c45b7d7f63de8bfeca8e55a5b3a1096dd3359a49a71273e20ab35c776b7dc5\": container with ID starting with c4c45b7d7f63de8bfeca8e55a5b3a1096dd3359a49a71273e20ab35c776b7dc5 not found: ID does not exist" containerID="c4c45b7d7f63de8bfeca8e55a5b3a1096dd3359a49a71273e20ab35c776b7dc5" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.059395 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4c45b7d7f63de8bfeca8e55a5b3a1096dd3359a49a71273e20ab35c776b7dc5"} err="failed to get container status \"c4c45b7d7f63de8bfeca8e55a5b3a1096dd3359a49a71273e20ab35c776b7dc5\": rpc error: code = NotFound desc = could not find container \"c4c45b7d7f63de8bfeca8e55a5b3a1096dd3359a49a71273e20ab35c776b7dc5\": container with ID starting with c4c45b7d7f63de8bfeca8e55a5b3a1096dd3359a49a71273e20ab35c776b7dc5 not found: ID does not exist" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.064623 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-config" (OuterVolumeSpecName: "config") pod "ee5d7a6d-2acf-4cfa-b028-de1c3a20c546" (UID: "ee5d7a6d-2acf-4cfa-b028-de1c3a20c546"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.070446 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="379fe74c-b561-44e6-8f06-2e0818922a09" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.236:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.070442 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="379fe74c-b561-44e6-8f06-2e0818922a09" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.236:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.072465 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ee5d7a6d-2acf-4cfa-b028-de1c3a20c546" (UID: "ee5d7a6d-2acf-4cfa-b028-de1c3a20c546"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.083649 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ee5d7a6d-2acf-4cfa-b028-de1c3a20c546" (UID: "ee5d7a6d-2acf-4cfa-b028-de1c3a20c546"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.091400 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpk9k\" (UniqueName: \"kubernetes.io/projected/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-kube-api-access-mpk9k\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.091424 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.091433 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.091441 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.091450 4934 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.091459 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:48 crc kubenswrapper[4934]: E1227 08:06:48.326763 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee5d7a6d_2acf_4cfa_b028_de1c3a20c546.slice/crio-ecaa5af55972576661cc763665718fa1c11cce6d34ed7ab546a465a3f6650dbc\": RecentStats: unable to find data in memory cache]" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.345136 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-l22fd"] Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.350860 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-l22fd"] Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.526604 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6kzb4" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.535110 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qxkzj" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.603493 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-scripts\") pod \"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e\" (UID: \"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e\") " Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.603549 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-combined-ca-bundle\") pod \"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e\" (UID: \"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e\") " Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.603694 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ct489\" (UniqueName: \"kubernetes.io/projected/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-kube-api-access-ct489\") pod \"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e\" (UID: \"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e\") " Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.603731 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-config-data\") pod \"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e\" (UID: \"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e\") " Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.608231 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-scripts" (OuterVolumeSpecName: "scripts") pod "52a3f24b-77b1-467e-b9ce-03e2ae5fe53e" (UID: "52a3f24b-77b1-467e-b9ce-03e2ae5fe53e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.609101 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-kube-api-access-ct489" (OuterVolumeSpecName: "kube-api-access-ct489") pod "52a3f24b-77b1-467e-b9ce-03e2ae5fe53e" (UID: "52a3f24b-77b1-467e-b9ce-03e2ae5fe53e"). InnerVolumeSpecName "kube-api-access-ct489". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.635074 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "52a3f24b-77b1-467e-b9ce-03e2ae5fe53e" (UID: "52a3f24b-77b1-467e-b9ce-03e2ae5fe53e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.667154 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-config-data" (OuterVolumeSpecName: "config-data") pod "52a3f24b-77b1-467e-b9ce-03e2ae5fe53e" (UID: "52a3f24b-77b1-467e-b9ce-03e2ae5fe53e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.705861 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-976zx\" (UniqueName: \"kubernetes.io/projected/94199ff4-bea2-4054-9752-7f9e5cd73d99-kube-api-access-976zx\") pod \"94199ff4-bea2-4054-9752-7f9e5cd73d99\" (UID: \"94199ff4-bea2-4054-9752-7f9e5cd73d99\") " Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.705932 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94199ff4-bea2-4054-9752-7f9e5cd73d99-combined-ca-bundle\") pod \"94199ff4-bea2-4054-9752-7f9e5cd73d99\" (UID: \"94199ff4-bea2-4054-9752-7f9e5cd73d99\") " Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.706033 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94199ff4-bea2-4054-9752-7f9e5cd73d99-scripts\") pod \"94199ff4-bea2-4054-9752-7f9e5cd73d99\" (UID: \"94199ff4-bea2-4054-9752-7f9e5cd73d99\") " Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.706062 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94199ff4-bea2-4054-9752-7f9e5cd73d99-config-data\") pod \"94199ff4-bea2-4054-9752-7f9e5cd73d99\" (UID: \"94199ff4-bea2-4054-9752-7f9e5cd73d99\") " Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.706739 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.706759 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.706769 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.706779 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ct489\" (UniqueName: \"kubernetes.io/projected/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e-kube-api-access-ct489\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.711275 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94199ff4-bea2-4054-9752-7f9e5cd73d99-kube-api-access-976zx" (OuterVolumeSpecName: "kube-api-access-976zx") pod "94199ff4-bea2-4054-9752-7f9e5cd73d99" (UID: "94199ff4-bea2-4054-9752-7f9e5cd73d99"). InnerVolumeSpecName "kube-api-access-976zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.722144 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94199ff4-bea2-4054-9752-7f9e5cd73d99-scripts" (OuterVolumeSpecName: "scripts") pod "94199ff4-bea2-4054-9752-7f9e5cd73d99" (UID: "94199ff4-bea2-4054-9752-7f9e5cd73d99"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.741068 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94199ff4-bea2-4054-9752-7f9e5cd73d99-config-data" (OuterVolumeSpecName: "config-data") pod "94199ff4-bea2-4054-9752-7f9e5cd73d99" (UID: "94199ff4-bea2-4054-9752-7f9e5cd73d99"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.751865 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94199ff4-bea2-4054-9752-7f9e5cd73d99-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94199ff4-bea2-4054-9752-7f9e5cd73d99" (UID: "94199ff4-bea2-4054-9752-7f9e5cd73d99"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.811011 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-976zx\" (UniqueName: \"kubernetes.io/projected/94199ff4-bea2-4054-9752-7f9e5cd73d99-kube-api-access-976zx\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.811063 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94199ff4-bea2-4054-9752-7f9e5cd73d99-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.811088 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94199ff4-bea2-4054-9752-7f9e5cd73d99-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.811127 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94199ff4-bea2-4054-9752-7f9e5cd73d99-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.975841 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-qxkzj" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.975849 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-qxkzj" event={"ID":"94199ff4-bea2-4054-9752-7f9e5cd73d99","Type":"ContainerDied","Data":"62afcb49a7b971c15c356e15a05d88355ab7581c5b339868724e2faacd2bc886"} Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.976390 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62afcb49a7b971c15c356e15a05d88355ab7581c5b339868724e2faacd2bc886" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.980720 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-6kzb4" event={"ID":"52a3f24b-77b1-467e-b9ce-03e2ae5fe53e","Type":"ContainerDied","Data":"9e0252c49143b7df4742f58d1f774094c51836957ffc5d715ccc0594ea80f313"} Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.980761 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e0252c49143b7df4742f58d1f774094c51836957ffc5d715ccc0594ea80f313" Dec 27 08:06:48 crc kubenswrapper[4934]: I1227 08:06:48.980807 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-6kzb4" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.075197 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 27 08:06:49 crc kubenswrapper[4934]: E1227 08:06:49.075920 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94199ff4-bea2-4054-9752-7f9e5cd73d99" containerName="nova-manage" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.075941 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="94199ff4-bea2-4054-9752-7f9e5cd73d99" containerName="nova-manage" Dec 27 08:06:49 crc kubenswrapper[4934]: E1227 08:06:49.075969 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee5d7a6d-2acf-4cfa-b028-de1c3a20c546" containerName="dnsmasq-dns" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.075978 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee5d7a6d-2acf-4cfa-b028-de1c3a20c546" containerName="dnsmasq-dns" Dec 27 08:06:49 crc kubenswrapper[4934]: E1227 08:06:49.076010 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee5d7a6d-2acf-4cfa-b028-de1c3a20c546" containerName="init" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.076019 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee5d7a6d-2acf-4cfa-b028-de1c3a20c546" containerName="init" Dec 27 08:06:49 crc kubenswrapper[4934]: E1227 08:06:49.076051 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72321394-c1c2-4254-9e2a-10617a3ba8c4" containerName="mariadb-account-create" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.076060 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="72321394-c1c2-4254-9e2a-10617a3ba8c4" containerName="mariadb-account-create" Dec 27 08:06:49 crc kubenswrapper[4934]: E1227 08:06:49.076077 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52a3f24b-77b1-467e-b9ce-03e2ae5fe53e" containerName="nova-cell1-conductor-db-sync" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.076088 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="52a3f24b-77b1-467e-b9ce-03e2ae5fe53e" containerName="nova-cell1-conductor-db-sync" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.076392 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee5d7a6d-2acf-4cfa-b028-de1c3a20c546" containerName="dnsmasq-dns" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.076413 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="72321394-c1c2-4254-9e2a-10617a3ba8c4" containerName="mariadb-account-create" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.076423 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="94199ff4-bea2-4054-9752-7f9e5cd73d99" containerName="nova-manage" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.076453 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="52a3f24b-77b1-467e-b9ce-03e2ae5fe53e" containerName="nova-cell1-conductor-db-sync" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.077665 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.080773 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.091259 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.187647 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-d4fxd"] Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.189360 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-d4fxd" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.194343 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.196730 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.197254 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-wrr58" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.210341 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-d4fxd"] Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.231076 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33eda6c6-6170-4293-83e0-ecce87ddca8c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"33eda6c6-6170-4293-83e0-ecce87ddca8c\") " pod="openstack/nova-cell1-conductor-0" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.231254 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbc79\" (UniqueName: \"kubernetes.io/projected/33eda6c6-6170-4293-83e0-ecce87ddca8c-kube-api-access-pbc79\") pod \"nova-cell1-conductor-0\" (UID: \"33eda6c6-6170-4293-83e0-ecce87ddca8c\") " pod="openstack/nova-cell1-conductor-0" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.231332 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33eda6c6-6170-4293-83e0-ecce87ddca8c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"33eda6c6-6170-4293-83e0-ecce87ddca8c\") " pod="openstack/nova-cell1-conductor-0" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.297508 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.298133 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="379fe74c-b561-44e6-8f06-2e0818922a09" containerName="nova-api-log" containerID="cri-o://b91ec4ad0a607ccb9f99f9f8a582817bcb77df25419ae2f0833daea69eb14b87" gracePeriod=30 Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.298311 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="379fe74c-b561-44e6-8f06-2e0818922a09" containerName="nova-api-api" containerID="cri-o://601d6356bc1737811808ec2361391369064bbccc90c3a004c42c79af9f7c5c98" gracePeriod=30 Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.307860 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.308034 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="794826da-01ec-4ee2-99dd-e5ece92df833" containerName="nova-scheduler-scheduler" containerID="cri-o://236fa6ccd4c0ae3a62409338ff24ba79be6a75b377d52ba6570f2da9651436b3" gracePeriod=30 Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.335693 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7jfd\" (UniqueName: \"kubernetes.io/projected/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-kube-api-access-j7jfd\") pod \"aodh-db-sync-d4fxd\" (UID: \"e83bd090-ad9e-4b6a-b72f-b6bdab47127a\") " pod="openstack/aodh-db-sync-d4fxd" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.335846 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33eda6c6-6170-4293-83e0-ecce87ddca8c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"33eda6c6-6170-4293-83e0-ecce87ddca8c\") " pod="openstack/nova-cell1-conductor-0" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.336191 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-config-data\") pod \"aodh-db-sync-d4fxd\" (UID: \"e83bd090-ad9e-4b6a-b72f-b6bdab47127a\") " pod="openstack/aodh-db-sync-d4fxd" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.336219 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-combined-ca-bundle\") pod \"aodh-db-sync-d4fxd\" (UID: \"e83bd090-ad9e-4b6a-b72f-b6bdab47127a\") " pod="openstack/aodh-db-sync-d4fxd" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.336276 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33eda6c6-6170-4293-83e0-ecce87ddca8c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"33eda6c6-6170-4293-83e0-ecce87ddca8c\") " pod="openstack/nova-cell1-conductor-0" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.336465 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-scripts\") pod \"aodh-db-sync-d4fxd\" (UID: \"e83bd090-ad9e-4b6a-b72f-b6bdab47127a\") " pod="openstack/aodh-db-sync-d4fxd" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.336565 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbc79\" (UniqueName: \"kubernetes.io/projected/33eda6c6-6170-4293-83e0-ecce87ddca8c-kube-api-access-pbc79\") pod \"nova-cell1-conductor-0\" (UID: \"33eda6c6-6170-4293-83e0-ecce87ddca8c\") " pod="openstack/nova-cell1-conductor-0" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.345214 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33eda6c6-6170-4293-83e0-ecce87ddca8c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"33eda6c6-6170-4293-83e0-ecce87ddca8c\") " pod="openstack/nova-cell1-conductor-0" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.345680 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33eda6c6-6170-4293-83e0-ecce87ddca8c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"33eda6c6-6170-4293-83e0-ecce87ddca8c\") " pod="openstack/nova-cell1-conductor-0" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.355711 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbc79\" (UniqueName: \"kubernetes.io/projected/33eda6c6-6170-4293-83e0-ecce87ddca8c-kube-api-access-pbc79\") pod \"nova-cell1-conductor-0\" (UID: \"33eda6c6-6170-4293-83e0-ecce87ddca8c\") " pod="openstack/nova-cell1-conductor-0" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.438453 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-config-data\") pod \"aodh-db-sync-d4fxd\" (UID: \"e83bd090-ad9e-4b6a-b72f-b6bdab47127a\") " pod="openstack/aodh-db-sync-d4fxd" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.438497 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-combined-ca-bundle\") pod \"aodh-db-sync-d4fxd\" (UID: \"e83bd090-ad9e-4b6a-b72f-b6bdab47127a\") " pod="openstack/aodh-db-sync-d4fxd" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.438584 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-scripts\") pod \"aodh-db-sync-d4fxd\" (UID: \"e83bd090-ad9e-4b6a-b72f-b6bdab47127a\") " pod="openstack/aodh-db-sync-d4fxd" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.438653 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7jfd\" (UniqueName: \"kubernetes.io/projected/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-kube-api-access-j7jfd\") pod \"aodh-db-sync-d4fxd\" (UID: \"e83bd090-ad9e-4b6a-b72f-b6bdab47127a\") " pod="openstack/aodh-db-sync-d4fxd" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.443632 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-combined-ca-bundle\") pod \"aodh-db-sync-d4fxd\" (UID: \"e83bd090-ad9e-4b6a-b72f-b6bdab47127a\") " pod="openstack/aodh-db-sync-d4fxd" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.444510 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-scripts\") pod \"aodh-db-sync-d4fxd\" (UID: \"e83bd090-ad9e-4b6a-b72f-b6bdab47127a\") " pod="openstack/aodh-db-sync-d4fxd" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.447199 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.447961 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-config-data\") pod \"aodh-db-sync-d4fxd\" (UID: \"e83bd090-ad9e-4b6a-b72f-b6bdab47127a\") " pod="openstack/aodh-db-sync-d4fxd" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.455770 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7jfd\" (UniqueName: \"kubernetes.io/projected/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-kube-api-access-j7jfd\") pod \"aodh-db-sync-d4fxd\" (UID: \"e83bd090-ad9e-4b6a-b72f-b6bdab47127a\") " pod="openstack/aodh-db-sync-d4fxd" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.490141 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee5d7a6d-2acf-4cfa-b028-de1c3a20c546" path="/var/lib/kubelet/pods/ee5d7a6d-2acf-4cfa-b028-de1c3a20c546/volumes" Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.515404 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-d4fxd" Dec 27 08:06:49 crc kubenswrapper[4934]: W1227 08:06:49.915831 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33eda6c6_6170_4293_83e0_ecce87ddca8c.slice/crio-0d5776ecdbab0dd917ad2233a6e6ab2cf5fc9c9c567db40f5ac5bb45109a742d WatchSource:0}: Error finding container 0d5776ecdbab0dd917ad2233a6e6ab2cf5fc9c9c567db40f5ac5bb45109a742d: Status 404 returned error can't find the container with id 0d5776ecdbab0dd917ad2233a6e6ab2cf5fc9c9c567db40f5ac5bb45109a742d Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.917222 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 27 08:06:49 crc kubenswrapper[4934]: I1227 08:06:49.996379 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"33eda6c6-6170-4293-83e0-ecce87ddca8c","Type":"ContainerStarted","Data":"0d5776ecdbab0dd917ad2233a6e6ab2cf5fc9c9c567db40f5ac5bb45109a742d"} Dec 27 08:06:50 crc kubenswrapper[4934]: I1227 08:06:50.001922 4934 generic.go:334] "Generic (PLEG): container finished" podID="379fe74c-b561-44e6-8f06-2e0818922a09" containerID="b91ec4ad0a607ccb9f99f9f8a582817bcb77df25419ae2f0833daea69eb14b87" exitCode=143 Dec 27 08:06:50 crc kubenswrapper[4934]: I1227 08:06:50.001959 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"379fe74c-b561-44e6-8f06-2e0818922a09","Type":"ContainerDied","Data":"b91ec4ad0a607ccb9f99f9f8a582817bcb77df25419ae2f0833daea69eb14b87"} Dec 27 08:06:50 crc kubenswrapper[4934]: I1227 08:06:50.173627 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-d4fxd"] Dec 27 08:06:51 crc kubenswrapper[4934]: I1227 08:06:51.018750 4934 generic.go:334] "Generic (PLEG): container finished" podID="794826da-01ec-4ee2-99dd-e5ece92df833" containerID="236fa6ccd4c0ae3a62409338ff24ba79be6a75b377d52ba6570f2da9651436b3" exitCode=0 Dec 27 08:06:51 crc kubenswrapper[4934]: I1227 08:06:51.019321 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"794826da-01ec-4ee2-99dd-e5ece92df833","Type":"ContainerDied","Data":"236fa6ccd4c0ae3a62409338ff24ba79be6a75b377d52ba6570f2da9651436b3"} Dec 27 08:06:51 crc kubenswrapper[4934]: I1227 08:06:51.019350 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"794826da-01ec-4ee2-99dd-e5ece92df833","Type":"ContainerDied","Data":"f152f356505fab18d34805b6c516f652628fa69640f40a31c4b9aab41b9159f2"} Dec 27 08:06:51 crc kubenswrapper[4934]: I1227 08:06:51.019363 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f152f356505fab18d34805b6c516f652628fa69640f40a31c4b9aab41b9159f2" Dec 27 08:06:51 crc kubenswrapper[4934]: I1227 08:06:51.021621 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"33eda6c6-6170-4293-83e0-ecce87ddca8c","Type":"ContainerStarted","Data":"831f7972032e12d8fc1673060475b7e3c35d476103b2afb5a8bad4b5c9fa2341"} Dec 27 08:06:51 crc kubenswrapper[4934]: I1227 08:06:51.022196 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 27 08:06:51 crc kubenswrapper[4934]: I1227 08:06:51.022981 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-d4fxd" event={"ID":"e83bd090-ad9e-4b6a-b72f-b6bdab47127a","Type":"ContainerStarted","Data":"b3824e62bcd2f8c0765c157c4d62f4769cfe1b68ba1e637c785b310ba4892a37"} Dec 27 08:06:51 crc kubenswrapper[4934]: I1227 08:06:51.043795 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.043205054 podStartE2EDuration="2.043205054s" podCreationTimestamp="2025-12-27 08:06:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:06:51.040166199 +0000 UTC m=+1471.860606793" watchObservedRunningTime="2025-12-27 08:06:51.043205054 +0000 UTC m=+1471.863645648" Dec 27 08:06:51 crc kubenswrapper[4934]: I1227 08:06:51.054541 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 27 08:06:51 crc kubenswrapper[4934]: I1227 08:06:51.204506 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nptb\" (UniqueName: \"kubernetes.io/projected/794826da-01ec-4ee2-99dd-e5ece92df833-kube-api-access-7nptb\") pod \"794826da-01ec-4ee2-99dd-e5ece92df833\" (UID: \"794826da-01ec-4ee2-99dd-e5ece92df833\") " Dec 27 08:06:51 crc kubenswrapper[4934]: I1227 08:06:51.204570 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/794826da-01ec-4ee2-99dd-e5ece92df833-config-data\") pod \"794826da-01ec-4ee2-99dd-e5ece92df833\" (UID: \"794826da-01ec-4ee2-99dd-e5ece92df833\") " Dec 27 08:06:51 crc kubenswrapper[4934]: I1227 08:06:51.204656 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/794826da-01ec-4ee2-99dd-e5ece92df833-combined-ca-bundle\") pod \"794826da-01ec-4ee2-99dd-e5ece92df833\" (UID: \"794826da-01ec-4ee2-99dd-e5ece92df833\") " Dec 27 08:06:51 crc kubenswrapper[4934]: I1227 08:06:51.216440 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/794826da-01ec-4ee2-99dd-e5ece92df833-kube-api-access-7nptb" (OuterVolumeSpecName: "kube-api-access-7nptb") pod "794826da-01ec-4ee2-99dd-e5ece92df833" (UID: "794826da-01ec-4ee2-99dd-e5ece92df833"). InnerVolumeSpecName "kube-api-access-7nptb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:06:51 crc kubenswrapper[4934]: I1227 08:06:51.245817 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/794826da-01ec-4ee2-99dd-e5ece92df833-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "794826da-01ec-4ee2-99dd-e5ece92df833" (UID: "794826da-01ec-4ee2-99dd-e5ece92df833"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:51 crc kubenswrapper[4934]: I1227 08:06:51.255888 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/794826da-01ec-4ee2-99dd-e5ece92df833-config-data" (OuterVolumeSpecName: "config-data") pod "794826da-01ec-4ee2-99dd-e5ece92df833" (UID: "794826da-01ec-4ee2-99dd-e5ece92df833"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:51 crc kubenswrapper[4934]: I1227 08:06:51.307301 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nptb\" (UniqueName: \"kubernetes.io/projected/794826da-01ec-4ee2-99dd-e5ece92df833-kube-api-access-7nptb\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:51 crc kubenswrapper[4934]: I1227 08:06:51.307337 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/794826da-01ec-4ee2-99dd-e5ece92df833-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:51 crc kubenswrapper[4934]: I1227 08:06:51.307351 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/794826da-01ec-4ee2-99dd-e5ece92df833-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:52 crc kubenswrapper[4934]: I1227 08:06:52.034173 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 27 08:06:52 crc kubenswrapper[4934]: I1227 08:06:52.063070 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 27 08:06:52 crc kubenswrapper[4934]: I1227 08:06:52.079362 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 27 08:06:52 crc kubenswrapper[4934]: I1227 08:06:52.093766 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 27 08:06:52 crc kubenswrapper[4934]: E1227 08:06:52.094392 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="794826da-01ec-4ee2-99dd-e5ece92df833" containerName="nova-scheduler-scheduler" Dec 27 08:06:52 crc kubenswrapper[4934]: I1227 08:06:52.094455 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="794826da-01ec-4ee2-99dd-e5ece92df833" containerName="nova-scheduler-scheduler" Dec 27 08:06:52 crc kubenswrapper[4934]: I1227 08:06:52.094777 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="794826da-01ec-4ee2-99dd-e5ece92df833" containerName="nova-scheduler-scheduler" Dec 27 08:06:52 crc kubenswrapper[4934]: I1227 08:06:52.095650 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 27 08:06:52 crc kubenswrapper[4934]: I1227 08:06:52.098308 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 27 08:06:52 crc kubenswrapper[4934]: I1227 08:06:52.119186 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 27 08:06:52 crc kubenswrapper[4934]: I1227 08:06:52.230277 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/933b14ea-5eba-4d47-a9b1-aead5e358154-config-data\") pod \"nova-scheduler-0\" (UID: \"933b14ea-5eba-4d47-a9b1-aead5e358154\") " pod="openstack/nova-scheduler-0" Dec 27 08:06:52 crc kubenswrapper[4934]: I1227 08:06:52.230405 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dtx5\" (UniqueName: \"kubernetes.io/projected/933b14ea-5eba-4d47-a9b1-aead5e358154-kube-api-access-2dtx5\") pod \"nova-scheduler-0\" (UID: \"933b14ea-5eba-4d47-a9b1-aead5e358154\") " pod="openstack/nova-scheduler-0" Dec 27 08:06:52 crc kubenswrapper[4934]: I1227 08:06:52.230738 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/933b14ea-5eba-4d47-a9b1-aead5e358154-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"933b14ea-5eba-4d47-a9b1-aead5e358154\") " pod="openstack/nova-scheduler-0" Dec 27 08:06:52 crc kubenswrapper[4934]: I1227 08:06:52.332825 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/933b14ea-5eba-4d47-a9b1-aead5e358154-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"933b14ea-5eba-4d47-a9b1-aead5e358154\") " pod="openstack/nova-scheduler-0" Dec 27 08:06:52 crc kubenswrapper[4934]: I1227 08:06:52.332923 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/933b14ea-5eba-4d47-a9b1-aead5e358154-config-data\") pod \"nova-scheduler-0\" (UID: \"933b14ea-5eba-4d47-a9b1-aead5e358154\") " pod="openstack/nova-scheduler-0" Dec 27 08:06:52 crc kubenswrapper[4934]: I1227 08:06:52.332975 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dtx5\" (UniqueName: \"kubernetes.io/projected/933b14ea-5eba-4d47-a9b1-aead5e358154-kube-api-access-2dtx5\") pod \"nova-scheduler-0\" (UID: \"933b14ea-5eba-4d47-a9b1-aead5e358154\") " pod="openstack/nova-scheduler-0" Dec 27 08:06:52 crc kubenswrapper[4934]: I1227 08:06:52.338674 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/933b14ea-5eba-4d47-a9b1-aead5e358154-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"933b14ea-5eba-4d47-a9b1-aead5e358154\") " pod="openstack/nova-scheduler-0" Dec 27 08:06:52 crc kubenswrapper[4934]: I1227 08:06:52.340857 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/933b14ea-5eba-4d47-a9b1-aead5e358154-config-data\") pod \"nova-scheduler-0\" (UID: \"933b14ea-5eba-4d47-a9b1-aead5e358154\") " pod="openstack/nova-scheduler-0" Dec 27 08:06:52 crc kubenswrapper[4934]: I1227 08:06:52.367421 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dtx5\" (UniqueName: \"kubernetes.io/projected/933b14ea-5eba-4d47-a9b1-aead5e358154-kube-api-access-2dtx5\") pod \"nova-scheduler-0\" (UID: \"933b14ea-5eba-4d47-a9b1-aead5e358154\") " pod="openstack/nova-scheduler-0" Dec 27 08:06:52 crc kubenswrapper[4934]: I1227 08:06:52.428167 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 27 08:06:52 crc kubenswrapper[4934]: I1227 08:06:52.869576 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-688b9f5b49-l22fd" podUID="ee5d7a6d-2acf-4cfa-b028-de1c3a20c546" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.211:5353: i/o timeout" Dec 27 08:06:53 crc kubenswrapper[4934]: I1227 08:06:53.483441 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="794826da-01ec-4ee2-99dd-e5ece92df833" path="/var/lib/kubelet/pods/794826da-01ec-4ee2-99dd-e5ece92df833/volumes" Dec 27 08:06:54 crc kubenswrapper[4934]: I1227 08:06:54.057358 4934 generic.go:334] "Generic (PLEG): container finished" podID="379fe74c-b561-44e6-8f06-2e0818922a09" containerID="601d6356bc1737811808ec2361391369064bbccc90c3a004c42c79af9f7c5c98" exitCode=0 Dec 27 08:06:54 crc kubenswrapper[4934]: I1227 08:06:54.057717 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"379fe74c-b561-44e6-8f06-2e0818922a09","Type":"ContainerDied","Data":"601d6356bc1737811808ec2361391369064bbccc90c3a004c42c79af9f7c5c98"} Dec 27 08:06:54 crc kubenswrapper[4934]: I1227 08:06:54.771042 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 27 08:06:54 crc kubenswrapper[4934]: I1227 08:06:54.909821 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/379fe74c-b561-44e6-8f06-2e0818922a09-config-data\") pod \"379fe74c-b561-44e6-8f06-2e0818922a09\" (UID: \"379fe74c-b561-44e6-8f06-2e0818922a09\") " Dec 27 08:06:54 crc kubenswrapper[4934]: I1227 08:06:54.909894 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-942r2\" (UniqueName: \"kubernetes.io/projected/379fe74c-b561-44e6-8f06-2e0818922a09-kube-api-access-942r2\") pod \"379fe74c-b561-44e6-8f06-2e0818922a09\" (UID: \"379fe74c-b561-44e6-8f06-2e0818922a09\") " Dec 27 08:06:54 crc kubenswrapper[4934]: I1227 08:06:54.909985 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/379fe74c-b561-44e6-8f06-2e0818922a09-combined-ca-bundle\") pod \"379fe74c-b561-44e6-8f06-2e0818922a09\" (UID: \"379fe74c-b561-44e6-8f06-2e0818922a09\") " Dec 27 08:06:54 crc kubenswrapper[4934]: I1227 08:06:54.910053 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/379fe74c-b561-44e6-8f06-2e0818922a09-logs\") pod \"379fe74c-b561-44e6-8f06-2e0818922a09\" (UID: \"379fe74c-b561-44e6-8f06-2e0818922a09\") " Dec 27 08:06:54 crc kubenswrapper[4934]: I1227 08:06:54.910583 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/379fe74c-b561-44e6-8f06-2e0818922a09-logs" (OuterVolumeSpecName: "logs") pod "379fe74c-b561-44e6-8f06-2e0818922a09" (UID: "379fe74c-b561-44e6-8f06-2e0818922a09"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:06:54 crc kubenswrapper[4934]: I1227 08:06:54.911410 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/379fe74c-b561-44e6-8f06-2e0818922a09-logs\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:54 crc kubenswrapper[4934]: I1227 08:06:54.914368 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/379fe74c-b561-44e6-8f06-2e0818922a09-kube-api-access-942r2" (OuterVolumeSpecName: "kube-api-access-942r2") pod "379fe74c-b561-44e6-8f06-2e0818922a09" (UID: "379fe74c-b561-44e6-8f06-2e0818922a09"). InnerVolumeSpecName "kube-api-access-942r2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:06:54 crc kubenswrapper[4934]: I1227 08:06:54.978159 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/379fe74c-b561-44e6-8f06-2e0818922a09-config-data" (OuterVolumeSpecName: "config-data") pod "379fe74c-b561-44e6-8f06-2e0818922a09" (UID: "379fe74c-b561-44e6-8f06-2e0818922a09"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:54 crc kubenswrapper[4934]: I1227 08:06:54.979712 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/379fe74c-b561-44e6-8f06-2e0818922a09-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "379fe74c-b561-44e6-8f06-2e0818922a09" (UID: "379fe74c-b561-44e6-8f06-2e0818922a09"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:55 crc kubenswrapper[4934]: W1227 08:06:55.012732 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod933b14ea_5eba_4d47_a9b1_aead5e358154.slice/crio-aaf175153bf72f40b031519732f510e2b908d9ff115e7866c43a6421eea177de WatchSource:0}: Error finding container aaf175153bf72f40b031519732f510e2b908d9ff115e7866c43a6421eea177de: Status 404 returned error can't find the container with id aaf175153bf72f40b031519732f510e2b908d9ff115e7866c43a6421eea177de Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.013986 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/379fe74c-b561-44e6-8f06-2e0818922a09-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.014696 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-942r2\" (UniqueName: \"kubernetes.io/projected/379fe74c-b561-44e6-8f06-2e0818922a09-kube-api-access-942r2\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.014718 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/379fe74c-b561-44e6-8f06-2e0818922a09-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.015832 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.074241 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-d4fxd" event={"ID":"e83bd090-ad9e-4b6a-b72f-b6bdab47127a","Type":"ContainerStarted","Data":"51323018708b2fb0bac023ea4fca5f9b434fc9be7f140175cd520dd278dc4336"} Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.078444 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"379fe74c-b561-44e6-8f06-2e0818922a09","Type":"ContainerDied","Data":"05df7fc936b43cab07fa0a1bdbb2fda7d3b7213529104e34be128889c453e9e9"} Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.078540 4934 scope.go:117] "RemoveContainer" containerID="601d6356bc1737811808ec2361391369064bbccc90c3a004c42c79af9f7c5c98" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.078698 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.083313 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"933b14ea-5eba-4d47-a9b1-aead5e358154","Type":"ContainerStarted","Data":"aaf175153bf72f40b031519732f510e2b908d9ff115e7866c43a6421eea177de"} Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.098814 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-d4fxd" podStartSLOduration=1.744096679 podStartE2EDuration="6.098794061s" podCreationTimestamp="2025-12-27 08:06:49 +0000 UTC" firstStartedPulling="2025-12-27 08:06:50.175661497 +0000 UTC m=+1470.996102091" lastFinishedPulling="2025-12-27 08:06:54.530358839 +0000 UTC m=+1475.350799473" observedRunningTime="2025-12-27 08:06:55.094341882 +0000 UTC m=+1475.914782476" watchObservedRunningTime="2025-12-27 08:06:55.098794061 +0000 UTC m=+1475.919234665" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.163165 4934 scope.go:117] "RemoveContainer" containerID="b91ec4ad0a607ccb9f99f9f8a582817bcb77df25419ae2f0833daea69eb14b87" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.196211 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.215988 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.238800 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 27 08:06:55 crc kubenswrapper[4934]: E1227 08:06:55.239356 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="379fe74c-b561-44e6-8f06-2e0818922a09" containerName="nova-api-log" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.239375 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="379fe74c-b561-44e6-8f06-2e0818922a09" containerName="nova-api-log" Dec 27 08:06:55 crc kubenswrapper[4934]: E1227 08:06:55.239397 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="379fe74c-b561-44e6-8f06-2e0818922a09" containerName="nova-api-api" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.239403 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="379fe74c-b561-44e6-8f06-2e0818922a09" containerName="nova-api-api" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.239631 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="379fe74c-b561-44e6-8f06-2e0818922a09" containerName="nova-api-log" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.239658 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="379fe74c-b561-44e6-8f06-2e0818922a09" containerName="nova-api-api" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.240947 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.244907 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.250047 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.424811 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9753e530-5b55-4951-af42-71297ce5c944-logs\") pod \"nova-api-0\" (UID: \"9753e530-5b55-4951-af42-71297ce5c944\") " pod="openstack/nova-api-0" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.425055 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9753e530-5b55-4951-af42-71297ce5c944-config-data\") pod \"nova-api-0\" (UID: \"9753e530-5b55-4951-af42-71297ce5c944\") " pod="openstack/nova-api-0" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.425159 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg5jp\" (UniqueName: \"kubernetes.io/projected/9753e530-5b55-4951-af42-71297ce5c944-kube-api-access-jg5jp\") pod \"nova-api-0\" (UID: \"9753e530-5b55-4951-af42-71297ce5c944\") " pod="openstack/nova-api-0" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.425496 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9753e530-5b55-4951-af42-71297ce5c944-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9753e530-5b55-4951-af42-71297ce5c944\") " pod="openstack/nova-api-0" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.483857 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="379fe74c-b561-44e6-8f06-2e0818922a09" path="/var/lib/kubelet/pods/379fe74c-b561-44e6-8f06-2e0818922a09/volumes" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.527882 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9753e530-5b55-4951-af42-71297ce5c944-logs\") pod \"nova-api-0\" (UID: \"9753e530-5b55-4951-af42-71297ce5c944\") " pod="openstack/nova-api-0" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.528302 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9753e530-5b55-4951-af42-71297ce5c944-logs\") pod \"nova-api-0\" (UID: \"9753e530-5b55-4951-af42-71297ce5c944\") " pod="openstack/nova-api-0" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.528484 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9753e530-5b55-4951-af42-71297ce5c944-config-data\") pod \"nova-api-0\" (UID: \"9753e530-5b55-4951-af42-71297ce5c944\") " pod="openstack/nova-api-0" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.528583 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg5jp\" (UniqueName: \"kubernetes.io/projected/9753e530-5b55-4951-af42-71297ce5c944-kube-api-access-jg5jp\") pod \"nova-api-0\" (UID: \"9753e530-5b55-4951-af42-71297ce5c944\") " pod="openstack/nova-api-0" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.529038 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9753e530-5b55-4951-af42-71297ce5c944-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9753e530-5b55-4951-af42-71297ce5c944\") " pod="openstack/nova-api-0" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.533833 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9753e530-5b55-4951-af42-71297ce5c944-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9753e530-5b55-4951-af42-71297ce5c944\") " pod="openstack/nova-api-0" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.545201 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9753e530-5b55-4951-af42-71297ce5c944-config-data\") pod \"nova-api-0\" (UID: \"9753e530-5b55-4951-af42-71297ce5c944\") " pod="openstack/nova-api-0" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.549733 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg5jp\" (UniqueName: \"kubernetes.io/projected/9753e530-5b55-4951-af42-71297ce5c944-kube-api-access-jg5jp\") pod \"nova-api-0\" (UID: \"9753e530-5b55-4951-af42-71297ce5c944\") " pod="openstack/nova-api-0" Dec 27 08:06:55 crc kubenswrapper[4934]: I1227 08:06:55.559483 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 27 08:06:56 crc kubenswrapper[4934]: W1227 08:06:56.051502 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9753e530_5b55_4951_af42_71297ce5c944.slice/crio-4e6425a3dd6c65706918495dc5d63673b54f7119128a75c36fe95438cbcbd365 WatchSource:0}: Error finding container 4e6425a3dd6c65706918495dc5d63673b54f7119128a75c36fe95438cbcbd365: Status 404 returned error can't find the container with id 4e6425a3dd6c65706918495dc5d63673b54f7119128a75c36fe95438cbcbd365 Dec 27 08:06:56 crc kubenswrapper[4934]: I1227 08:06:56.052049 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 27 08:06:56 crc kubenswrapper[4934]: I1227 08:06:56.099844 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9753e530-5b55-4951-af42-71297ce5c944","Type":"ContainerStarted","Data":"4e6425a3dd6c65706918495dc5d63673b54f7119128a75c36fe95438cbcbd365"} Dec 27 08:06:56 crc kubenswrapper[4934]: I1227 08:06:56.106331 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"933b14ea-5eba-4d47-a9b1-aead5e358154","Type":"ContainerStarted","Data":"a8860841c6945191b9104dd6297726f8026d1078a711851ae9e09c7c9c5015c7"} Dec 27 08:06:56 crc kubenswrapper[4934]: I1227 08:06:56.135037 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=4.135017747 podStartE2EDuration="4.135017747s" podCreationTimestamp="2025-12-27 08:06:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:06:56.129123791 +0000 UTC m=+1476.949564385" watchObservedRunningTime="2025-12-27 08:06:56.135017747 +0000 UTC m=+1476.955458351" Dec 27 08:06:57 crc kubenswrapper[4934]: I1227 08:06:57.118353 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9753e530-5b55-4951-af42-71297ce5c944","Type":"ContainerStarted","Data":"a049ffa669f4d9521ebf64ba2c09b7a1ac770f1963ed43a509e333623eca0f4e"} Dec 27 08:06:57 crc kubenswrapper[4934]: I1227 08:06:57.118652 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9753e530-5b55-4951-af42-71297ce5c944","Type":"ContainerStarted","Data":"f8f94b18eff1c88a06c4d34a79292ae09139ccfbc7cd8e5fae2b49b1571e74e8"} Dec 27 08:06:57 crc kubenswrapper[4934]: I1227 08:06:57.137655 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.137636524 podStartE2EDuration="2.137636524s" podCreationTimestamp="2025-12-27 08:06:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:06:57.133649495 +0000 UTC m=+1477.954090099" watchObservedRunningTime="2025-12-27 08:06:57.137636524 +0000 UTC m=+1477.958077118" Dec 27 08:06:57 crc kubenswrapper[4934]: I1227 08:06:57.429419 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 27 08:06:58 crc kubenswrapper[4934]: I1227 08:06:58.130076 4934 generic.go:334] "Generic (PLEG): container finished" podID="e83bd090-ad9e-4b6a-b72f-b6bdab47127a" containerID="51323018708b2fb0bac023ea4fca5f9b434fc9be7f140175cd520dd278dc4336" exitCode=0 Dec 27 08:06:58 crc kubenswrapper[4934]: I1227 08:06:58.130216 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-d4fxd" event={"ID":"e83bd090-ad9e-4b6a-b72f-b6bdab47127a","Type":"ContainerDied","Data":"51323018708b2fb0bac023ea4fca5f9b434fc9be7f140175cd520dd278dc4336"} Dec 27 08:06:59 crc kubenswrapper[4934]: I1227 08:06:59.489148 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 27 08:06:59 crc kubenswrapper[4934]: I1227 08:06:59.656161 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-d4fxd" Dec 27 08:06:59 crc kubenswrapper[4934]: I1227 08:06:59.765440 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-scripts\") pod \"e83bd090-ad9e-4b6a-b72f-b6bdab47127a\" (UID: \"e83bd090-ad9e-4b6a-b72f-b6bdab47127a\") " Dec 27 08:06:59 crc kubenswrapper[4934]: I1227 08:06:59.765573 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-combined-ca-bundle\") pod \"e83bd090-ad9e-4b6a-b72f-b6bdab47127a\" (UID: \"e83bd090-ad9e-4b6a-b72f-b6bdab47127a\") " Dec 27 08:06:59 crc kubenswrapper[4934]: I1227 08:06:59.765639 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7jfd\" (UniqueName: \"kubernetes.io/projected/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-kube-api-access-j7jfd\") pod \"e83bd090-ad9e-4b6a-b72f-b6bdab47127a\" (UID: \"e83bd090-ad9e-4b6a-b72f-b6bdab47127a\") " Dec 27 08:06:59 crc kubenswrapper[4934]: I1227 08:06:59.765833 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-config-data\") pod \"e83bd090-ad9e-4b6a-b72f-b6bdab47127a\" (UID: \"e83bd090-ad9e-4b6a-b72f-b6bdab47127a\") " Dec 27 08:06:59 crc kubenswrapper[4934]: I1227 08:06:59.770959 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-scripts" (OuterVolumeSpecName: "scripts") pod "e83bd090-ad9e-4b6a-b72f-b6bdab47127a" (UID: "e83bd090-ad9e-4b6a-b72f-b6bdab47127a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:59 crc kubenswrapper[4934]: I1227 08:06:59.771557 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-kube-api-access-j7jfd" (OuterVolumeSpecName: "kube-api-access-j7jfd") pod "e83bd090-ad9e-4b6a-b72f-b6bdab47127a" (UID: "e83bd090-ad9e-4b6a-b72f-b6bdab47127a"). InnerVolumeSpecName "kube-api-access-j7jfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:06:59 crc kubenswrapper[4934]: I1227 08:06:59.811725 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-config-data" (OuterVolumeSpecName: "config-data") pod "e83bd090-ad9e-4b6a-b72f-b6bdab47127a" (UID: "e83bd090-ad9e-4b6a-b72f-b6bdab47127a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:59 crc kubenswrapper[4934]: I1227 08:06:59.830365 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e83bd090-ad9e-4b6a-b72f-b6bdab47127a" (UID: "e83bd090-ad9e-4b6a-b72f-b6bdab47127a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:06:59 crc kubenswrapper[4934]: I1227 08:06:59.869873 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7jfd\" (UniqueName: \"kubernetes.io/projected/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-kube-api-access-j7jfd\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:59 crc kubenswrapper[4934]: I1227 08:06:59.869911 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:59 crc kubenswrapper[4934]: I1227 08:06:59.869928 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:06:59 crc kubenswrapper[4934]: I1227 08:06:59.869969 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e83bd090-ad9e-4b6a-b72f-b6bdab47127a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:00 crc kubenswrapper[4934]: I1227 08:07:00.151337 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-d4fxd" event={"ID":"e83bd090-ad9e-4b6a-b72f-b6bdab47127a","Type":"ContainerDied","Data":"b3824e62bcd2f8c0765c157c4d62f4769cfe1b68ba1e637c785b310ba4892a37"} Dec 27 08:07:00 crc kubenswrapper[4934]: I1227 08:07:00.151664 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3824e62bcd2f8c0765c157c4d62f4769cfe1b68ba1e637c785b310ba4892a37" Dec 27 08:07:00 crc kubenswrapper[4934]: I1227 08:07:00.151720 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-d4fxd" Dec 27 08:07:01 crc kubenswrapper[4934]: I1227 08:07:01.824792 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 27 08:07:02 crc kubenswrapper[4934]: I1227 08:07:02.429417 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 27 08:07:02 crc kubenswrapper[4934]: I1227 08:07:02.485397 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 27 08:07:03 crc kubenswrapper[4934]: I1227 08:07:03.249895 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 27 08:07:03 crc kubenswrapper[4934]: I1227 08:07:03.708380 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 27 08:07:03 crc kubenswrapper[4934]: E1227 08:07:03.708961 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e83bd090-ad9e-4b6a-b72f-b6bdab47127a" containerName="aodh-db-sync" Dec 27 08:07:03 crc kubenswrapper[4934]: I1227 08:07:03.708979 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="e83bd090-ad9e-4b6a-b72f-b6bdab47127a" containerName="aodh-db-sync" Dec 27 08:07:03 crc kubenswrapper[4934]: I1227 08:07:03.709192 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="e83bd090-ad9e-4b6a-b72f-b6bdab47127a" containerName="aodh-db-sync" Dec 27 08:07:03 crc kubenswrapper[4934]: I1227 08:07:03.711244 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 27 08:07:03 crc kubenswrapper[4934]: I1227 08:07:03.714376 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 27 08:07:03 crc kubenswrapper[4934]: I1227 08:07:03.714497 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 27 08:07:03 crc kubenswrapper[4934]: I1227 08:07:03.714883 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-wrr58" Dec 27 08:07:03 crc kubenswrapper[4934]: I1227 08:07:03.732114 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 27 08:07:03 crc kubenswrapper[4934]: I1227 08:07:03.879145 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlqmv\" (UniqueName: \"kubernetes.io/projected/1d13d0a6-ede9-45fe-9338-5a90260f0455-kube-api-access-mlqmv\") pod \"aodh-0\" (UID: \"1d13d0a6-ede9-45fe-9338-5a90260f0455\") " pod="openstack/aodh-0" Dec 27 08:07:03 crc kubenswrapper[4934]: I1227 08:07:03.879279 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d13d0a6-ede9-45fe-9338-5a90260f0455-config-data\") pod \"aodh-0\" (UID: \"1d13d0a6-ede9-45fe-9338-5a90260f0455\") " pod="openstack/aodh-0" Dec 27 08:07:03 crc kubenswrapper[4934]: I1227 08:07:03.879381 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d13d0a6-ede9-45fe-9338-5a90260f0455-combined-ca-bundle\") pod \"aodh-0\" (UID: \"1d13d0a6-ede9-45fe-9338-5a90260f0455\") " pod="openstack/aodh-0" Dec 27 08:07:03 crc kubenswrapper[4934]: I1227 08:07:03.879470 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d13d0a6-ede9-45fe-9338-5a90260f0455-scripts\") pod \"aodh-0\" (UID: \"1d13d0a6-ede9-45fe-9338-5a90260f0455\") " pod="openstack/aodh-0" Dec 27 08:07:03 crc kubenswrapper[4934]: I1227 08:07:03.981479 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d13d0a6-ede9-45fe-9338-5a90260f0455-combined-ca-bundle\") pod \"aodh-0\" (UID: \"1d13d0a6-ede9-45fe-9338-5a90260f0455\") " pod="openstack/aodh-0" Dec 27 08:07:03 crc kubenswrapper[4934]: I1227 08:07:03.981752 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d13d0a6-ede9-45fe-9338-5a90260f0455-scripts\") pod \"aodh-0\" (UID: \"1d13d0a6-ede9-45fe-9338-5a90260f0455\") " pod="openstack/aodh-0" Dec 27 08:07:03 crc kubenswrapper[4934]: I1227 08:07:03.981824 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlqmv\" (UniqueName: \"kubernetes.io/projected/1d13d0a6-ede9-45fe-9338-5a90260f0455-kube-api-access-mlqmv\") pod \"aodh-0\" (UID: \"1d13d0a6-ede9-45fe-9338-5a90260f0455\") " pod="openstack/aodh-0" Dec 27 08:07:03 crc kubenswrapper[4934]: I1227 08:07:03.981942 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d13d0a6-ede9-45fe-9338-5a90260f0455-config-data\") pod \"aodh-0\" (UID: \"1d13d0a6-ede9-45fe-9338-5a90260f0455\") " pod="openstack/aodh-0" Dec 27 08:07:03 crc kubenswrapper[4934]: I1227 08:07:03.988507 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d13d0a6-ede9-45fe-9338-5a90260f0455-scripts\") pod \"aodh-0\" (UID: \"1d13d0a6-ede9-45fe-9338-5a90260f0455\") " pod="openstack/aodh-0" Dec 27 08:07:03 crc kubenswrapper[4934]: I1227 08:07:03.993861 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d13d0a6-ede9-45fe-9338-5a90260f0455-combined-ca-bundle\") pod \"aodh-0\" (UID: \"1d13d0a6-ede9-45fe-9338-5a90260f0455\") " pod="openstack/aodh-0" Dec 27 08:07:03 crc kubenswrapper[4934]: I1227 08:07:03.994274 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d13d0a6-ede9-45fe-9338-5a90260f0455-config-data\") pod \"aodh-0\" (UID: \"1d13d0a6-ede9-45fe-9338-5a90260f0455\") " pod="openstack/aodh-0" Dec 27 08:07:04 crc kubenswrapper[4934]: I1227 08:07:04.008614 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlqmv\" (UniqueName: \"kubernetes.io/projected/1d13d0a6-ede9-45fe-9338-5a90260f0455-kube-api-access-mlqmv\") pod \"aodh-0\" (UID: \"1d13d0a6-ede9-45fe-9338-5a90260f0455\") " pod="openstack/aodh-0" Dec 27 08:07:04 crc kubenswrapper[4934]: I1227 08:07:04.041907 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 27 08:07:04 crc kubenswrapper[4934]: I1227 08:07:04.555038 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 27 08:07:05 crc kubenswrapper[4934]: I1227 08:07:05.237778 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1d13d0a6-ede9-45fe-9338-5a90260f0455","Type":"ContainerStarted","Data":"de00a807aca97606daac3aa01aa6491dca54b465662476046c10287a682e0bf9"} Dec 27 08:07:05 crc kubenswrapper[4934]: I1227 08:07:05.559953 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 27 08:07:05 crc kubenswrapper[4934]: I1227 08:07:05.560016 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 27 08:07:05 crc kubenswrapper[4934]: I1227 08:07:05.957157 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:07:05 crc kubenswrapper[4934]: I1227 08:07:05.957667 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="600b5343-a21b-4c61-94b4-54678000d8de" containerName="ceilometer-central-agent" containerID="cri-o://6dcda01dab092e5244c68ae57ba68aabf16cff612d634574698d3e67e2504305" gracePeriod=30 Dec 27 08:07:05 crc kubenswrapper[4934]: I1227 08:07:05.957729 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="600b5343-a21b-4c61-94b4-54678000d8de" containerName="proxy-httpd" containerID="cri-o://9674809e36b8880f24e5e524b5c8bd1353ae482612663892a3e294244f5de59f" gracePeriod=30 Dec 27 08:07:05 crc kubenswrapper[4934]: I1227 08:07:05.957785 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="600b5343-a21b-4c61-94b4-54678000d8de" containerName="sg-core" containerID="cri-o://36e64ee6387a3cd1b80d20e2a97954d73739eac60a5f9361a858c4a90b0fdcd5" gracePeriod=30 Dec 27 08:07:05 crc kubenswrapper[4934]: I1227 08:07:05.957822 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="600b5343-a21b-4c61-94b4-54678000d8de" containerName="ceilometer-notification-agent" containerID="cri-o://7c39932385049c834e0a7ae64966f692854ef783b5795bf025e9f64f9eafe5eb" gracePeriod=30 Dec 27 08:07:06 crc kubenswrapper[4934]: I1227 08:07:06.256134 4934 generic.go:334] "Generic (PLEG): container finished" podID="600b5343-a21b-4c61-94b4-54678000d8de" containerID="36e64ee6387a3cd1b80d20e2a97954d73739eac60a5f9361a858c4a90b0fdcd5" exitCode=2 Dec 27 08:07:06 crc kubenswrapper[4934]: I1227 08:07:06.256205 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"600b5343-a21b-4c61-94b4-54678000d8de","Type":"ContainerDied","Data":"36e64ee6387a3cd1b80d20e2a97954d73739eac60a5f9361a858c4a90b0fdcd5"} Dec 27 08:07:06 crc kubenswrapper[4934]: I1227 08:07:06.260520 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1d13d0a6-ede9-45fe-9338-5a90260f0455","Type":"ContainerStarted","Data":"fd596c2c4a357feb978a9acf180d8d3ae8340b1b043c17107ff33fe257d0ce7d"} Dec 27 08:07:06 crc kubenswrapper[4934]: I1227 08:07:06.643266 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9753e530-5b55-4951-af42-71297ce5c944" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.244:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 08:07:06 crc kubenswrapper[4934]: I1227 08:07:06.643353 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9753e530-5b55-4951-af42-71297ce5c944" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.244:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.284435 4934 generic.go:334] "Generic (PLEG): container finished" podID="600b5343-a21b-4c61-94b4-54678000d8de" containerID="9674809e36b8880f24e5e524b5c8bd1353ae482612663892a3e294244f5de59f" exitCode=0 Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.284814 4934 generic.go:334] "Generic (PLEG): container finished" podID="600b5343-a21b-4c61-94b4-54678000d8de" containerID="7c39932385049c834e0a7ae64966f692854ef783b5795bf025e9f64f9eafe5eb" exitCode=0 Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.284828 4934 generic.go:334] "Generic (PLEG): container finished" podID="600b5343-a21b-4c61-94b4-54678000d8de" containerID="6dcda01dab092e5244c68ae57ba68aabf16cff612d634574698d3e67e2504305" exitCode=0 Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.284885 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"600b5343-a21b-4c61-94b4-54678000d8de","Type":"ContainerDied","Data":"9674809e36b8880f24e5e524b5c8bd1353ae482612663892a3e294244f5de59f"} Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.284915 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"600b5343-a21b-4c61-94b4-54678000d8de","Type":"ContainerDied","Data":"7c39932385049c834e0a7ae64966f692854ef783b5795bf025e9f64f9eafe5eb"} Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.284930 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"600b5343-a21b-4c61-94b4-54678000d8de","Type":"ContainerDied","Data":"6dcda01dab092e5244c68ae57ba68aabf16cff612d634574698d3e67e2504305"} Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.289134 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1d13d0a6-ede9-45fe-9338-5a90260f0455","Type":"ContainerStarted","Data":"3acc162dd29cb480c5e3a7b72d533b66a8a7fa5367914e43729749c74ad05402"} Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.410682 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.566707 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/600b5343-a21b-4c61-94b4-54678000d8de-log-httpd\") pod \"600b5343-a21b-4c61-94b4-54678000d8de\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.566797 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-combined-ca-bundle\") pod \"600b5343-a21b-4c61-94b4-54678000d8de\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.566992 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-scripts\") pod \"600b5343-a21b-4c61-94b4-54678000d8de\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.567130 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/600b5343-a21b-4c61-94b4-54678000d8de-run-httpd\") pod \"600b5343-a21b-4c61-94b4-54678000d8de\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.567172 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-config-data\") pod \"600b5343-a21b-4c61-94b4-54678000d8de\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.567210 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nhk4\" (UniqueName: \"kubernetes.io/projected/600b5343-a21b-4c61-94b4-54678000d8de-kube-api-access-7nhk4\") pod \"600b5343-a21b-4c61-94b4-54678000d8de\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.567301 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-sg-core-conf-yaml\") pod \"600b5343-a21b-4c61-94b4-54678000d8de\" (UID: \"600b5343-a21b-4c61-94b4-54678000d8de\") " Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.567313 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/600b5343-a21b-4c61-94b4-54678000d8de-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "600b5343-a21b-4c61-94b4-54678000d8de" (UID: "600b5343-a21b-4c61-94b4-54678000d8de"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.567327 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/600b5343-a21b-4c61-94b4-54678000d8de-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "600b5343-a21b-4c61-94b4-54678000d8de" (UID: "600b5343-a21b-4c61-94b4-54678000d8de"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.569009 4934 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/600b5343-a21b-4c61-94b4-54678000d8de-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.569152 4934 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/600b5343-a21b-4c61-94b4-54678000d8de-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.573111 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/600b5343-a21b-4c61-94b4-54678000d8de-kube-api-access-7nhk4" (OuterVolumeSpecName: "kube-api-access-7nhk4") pod "600b5343-a21b-4c61-94b4-54678000d8de" (UID: "600b5343-a21b-4c61-94b4-54678000d8de"). InnerVolumeSpecName "kube-api-access-7nhk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.577240 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-scripts" (OuterVolumeSpecName: "scripts") pod "600b5343-a21b-4c61-94b4-54678000d8de" (UID: "600b5343-a21b-4c61-94b4-54678000d8de"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.620161 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "600b5343-a21b-4c61-94b4-54678000d8de" (UID: "600b5343-a21b-4c61-94b4-54678000d8de"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.661379 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "600b5343-a21b-4c61-94b4-54678000d8de" (UID: "600b5343-a21b-4c61-94b4-54678000d8de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.676991 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nhk4\" (UniqueName: \"kubernetes.io/projected/600b5343-a21b-4c61-94b4-54678000d8de-kube-api-access-7nhk4\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.677032 4934 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.677047 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.677062 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.714598 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-config-data" (OuterVolumeSpecName: "config-data") pod "600b5343-a21b-4c61-94b4-54678000d8de" (UID: "600b5343-a21b-4c61-94b4-54678000d8de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:07 crc kubenswrapper[4934]: I1227 08:07:07.779468 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/600b5343-a21b-4c61-94b4-54678000d8de-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.306843 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"600b5343-a21b-4c61-94b4-54678000d8de","Type":"ContainerDied","Data":"3d1805306412f0394ea8cde2cf554718742f81a85edf74294e9f4762e1b653c0"} Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.307136 4934 scope.go:117] "RemoveContainer" containerID="9674809e36b8880f24e5e524b5c8bd1353ae482612663892a3e294244f5de59f" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.306925 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.374167 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.393215 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.404379 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:07:08 crc kubenswrapper[4934]: E1227 08:07:08.404933 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="600b5343-a21b-4c61-94b4-54678000d8de" containerName="ceilometer-central-agent" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.404952 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="600b5343-a21b-4c61-94b4-54678000d8de" containerName="ceilometer-central-agent" Dec 27 08:07:08 crc kubenswrapper[4934]: E1227 08:07:08.404987 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="600b5343-a21b-4c61-94b4-54678000d8de" containerName="sg-core" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.404996 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="600b5343-a21b-4c61-94b4-54678000d8de" containerName="sg-core" Dec 27 08:07:08 crc kubenswrapper[4934]: E1227 08:07:08.405006 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="600b5343-a21b-4c61-94b4-54678000d8de" containerName="proxy-httpd" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.405011 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="600b5343-a21b-4c61-94b4-54678000d8de" containerName="proxy-httpd" Dec 27 08:07:08 crc kubenswrapper[4934]: E1227 08:07:08.405019 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="600b5343-a21b-4c61-94b4-54678000d8de" containerName="ceilometer-notification-agent" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.405024 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="600b5343-a21b-4c61-94b4-54678000d8de" containerName="ceilometer-notification-agent" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.405249 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="600b5343-a21b-4c61-94b4-54678000d8de" containerName="proxy-httpd" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.405264 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="600b5343-a21b-4c61-94b4-54678000d8de" containerName="sg-core" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.405272 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="600b5343-a21b-4c61-94b4-54678000d8de" containerName="ceilometer-central-agent" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.405295 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="600b5343-a21b-4c61-94b4-54678000d8de" containerName="ceilometer-notification-agent" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.408522 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.412317 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.412499 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.421060 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.497783 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-log-httpd\") pod \"ceilometer-0\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.497852 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frmkn\" (UniqueName: \"kubernetes.io/projected/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-kube-api-access-frmkn\") pod \"ceilometer-0\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.497883 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-run-httpd\") pod \"ceilometer-0\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.497928 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-scripts\") pod \"ceilometer-0\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.497959 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-config-data\") pod \"ceilometer-0\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.498170 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.498290 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.600467 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.600598 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-log-httpd\") pod \"ceilometer-0\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.600639 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frmkn\" (UniqueName: \"kubernetes.io/projected/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-kube-api-access-frmkn\") pod \"ceilometer-0\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.600665 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-run-httpd\") pod \"ceilometer-0\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.600687 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-scripts\") pod \"ceilometer-0\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.600724 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-config-data\") pod \"ceilometer-0\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.600928 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.601449 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-run-httpd\") pod \"ceilometer-0\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.601610 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-log-httpd\") pod \"ceilometer-0\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.605592 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-scripts\") pod \"ceilometer-0\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.612942 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.615102 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-config-data\") pod \"ceilometer-0\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.617392 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.618291 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frmkn\" (UniqueName: \"kubernetes.io/projected/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-kube-api-access-frmkn\") pod \"ceilometer-0\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.724955 4934 scope.go:117] "RemoveContainer" containerID="36e64ee6387a3cd1b80d20e2a97954d73739eac60a5f9361a858c4a90b0fdcd5" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.731857 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.768799 4934 scope.go:117] "RemoveContainer" containerID="7c39932385049c834e0a7ae64966f692854ef783b5795bf025e9f64f9eafe5eb" Dec 27 08:07:08 crc kubenswrapper[4934]: I1227 08:07:08.909000 4934 scope.go:117] "RemoveContainer" containerID="6dcda01dab092e5244c68ae57ba68aabf16cff612d634574698d3e67e2504305" Dec 27 08:07:09 crc kubenswrapper[4934]: I1227 08:07:09.015780 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 27 08:07:09 crc kubenswrapper[4934]: I1227 08:07:09.321420 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1d13d0a6-ede9-45fe-9338-5a90260f0455","Type":"ContainerStarted","Data":"e6cf52ef3f84a1b3eede7e7564cb23f48e1ed825f07acb54857b63501d13c82a"} Dec 27 08:07:09 crc kubenswrapper[4934]: I1227 08:07:09.384558 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:07:09 crc kubenswrapper[4934]: I1227 08:07:09.480551 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="600b5343-a21b-4c61-94b4-54678000d8de" path="/var/lib/kubelet/pods/600b5343-a21b-4c61-94b4-54678000d8de/volumes" Dec 27 08:07:10 crc kubenswrapper[4934]: I1227 08:07:10.336558 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2fad2302-c4ec-4c95-a25e-e0ef03a99b61","Type":"ContainerStarted","Data":"75ebb1cbce396521261708603d5b5be37fc812e042adf87066ab5b10f460be16"} Dec 27 08:07:10 crc kubenswrapper[4934]: I1227 08:07:10.570832 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:07:12 crc kubenswrapper[4934]: I1227 08:07:12.358286 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2fad2302-c4ec-4c95-a25e-e0ef03a99b61","Type":"ContainerStarted","Data":"3fa1c7010492f1fe94460776a4a760f0d0e68513d6db66ec3b278d16f660996f"} Dec 27 08:07:12 crc kubenswrapper[4934]: I1227 08:07:12.364618 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1d13d0a6-ede9-45fe-9338-5a90260f0455","Type":"ContainerStarted","Data":"9ea1535deb217127b7d666092376b7a0856f76af6c706047e65a73355eafb6e4"} Dec 27 08:07:12 crc kubenswrapper[4934]: I1227 08:07:12.364863 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="1d13d0a6-ede9-45fe-9338-5a90260f0455" containerName="aodh-api" containerID="cri-o://fd596c2c4a357feb978a9acf180d8d3ae8340b1b043c17107ff33fe257d0ce7d" gracePeriod=30 Dec 27 08:07:12 crc kubenswrapper[4934]: I1227 08:07:12.364987 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="1d13d0a6-ede9-45fe-9338-5a90260f0455" containerName="aodh-notifier" containerID="cri-o://e6cf52ef3f84a1b3eede7e7564cb23f48e1ed825f07acb54857b63501d13c82a" gracePeriod=30 Dec 27 08:07:12 crc kubenswrapper[4934]: I1227 08:07:12.365044 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="1d13d0a6-ede9-45fe-9338-5a90260f0455" containerName="aodh-evaluator" containerID="cri-o://3acc162dd29cb480c5e3a7b72d533b66a8a7fa5367914e43729749c74ad05402" gracePeriod=30 Dec 27 08:07:12 crc kubenswrapper[4934]: I1227 08:07:12.369710 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="1d13d0a6-ede9-45fe-9338-5a90260f0455" containerName="aodh-listener" containerID="cri-o://9ea1535deb217127b7d666092376b7a0856f76af6c706047e65a73355eafb6e4" gracePeriod=30 Dec 27 08:07:12 crc kubenswrapper[4934]: I1227 08:07:12.407357 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=3.544031073 podStartE2EDuration="9.407332955s" podCreationTimestamp="2025-12-27 08:07:03 +0000 UTC" firstStartedPulling="2025-12-27 08:07:04.561458204 +0000 UTC m=+1485.381898798" lastFinishedPulling="2025-12-27 08:07:10.424760086 +0000 UTC m=+1491.245200680" observedRunningTime="2025-12-27 08:07:12.390504428 +0000 UTC m=+1493.210945022" watchObservedRunningTime="2025-12-27 08:07:12.407332955 +0000 UTC m=+1493.227773559" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.312687 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.380435 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2fad2302-c4ec-4c95-a25e-e0ef03a99b61","Type":"ContainerStarted","Data":"a1159f38cea2415aafb205147cca11c7cff559a38e5b1821fbfe82ff2e7fee1f"} Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.387190 4934 generic.go:334] "Generic (PLEG): container finished" podID="0f0f40ab-1168-4fea-84b1-b134eb2a1b3c" containerID="bc6848ec46001daa30195a2f745ad7629c7bf68dd56b59d8b02fda09322ce597" exitCode=137 Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.387306 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.387852 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0f0f40ab-1168-4fea-84b1-b134eb2a1b3c","Type":"ContainerDied","Data":"bc6848ec46001daa30195a2f745ad7629c7bf68dd56b59d8b02fda09322ce597"} Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.387885 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"0f0f40ab-1168-4fea-84b1-b134eb2a1b3c","Type":"ContainerDied","Data":"d3712aea3b2913aa199387cdf9f87000c6483d558814a9708bc3320132fb6272"} Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.387907 4934 scope.go:117] "RemoveContainer" containerID="bc6848ec46001daa30195a2f745ad7629c7bf68dd56b59d8b02fda09322ce597" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.390292 4934 generic.go:334] "Generic (PLEG): container finished" podID="4c696b92-7128-4465-875f-d2ef4298e798" containerID="bc9b8337126f66220fa9bcfe2a59c8bf4423fcc2292e6bbc9bf9ca725a983c0b" exitCode=137 Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.390370 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c696b92-7128-4465-875f-d2ef4298e798","Type":"ContainerDied","Data":"bc9b8337126f66220fa9bcfe2a59c8bf4423fcc2292e6bbc9bf9ca725a983c0b"} Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.407266 4934 generic.go:334] "Generic (PLEG): container finished" podID="1d13d0a6-ede9-45fe-9338-5a90260f0455" containerID="e6cf52ef3f84a1b3eede7e7564cb23f48e1ed825f07acb54857b63501d13c82a" exitCode=0 Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.407307 4934 generic.go:334] "Generic (PLEG): container finished" podID="1d13d0a6-ede9-45fe-9338-5a90260f0455" containerID="3acc162dd29cb480c5e3a7b72d533b66a8a7fa5367914e43729749c74ad05402" exitCode=0 Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.407317 4934 generic.go:334] "Generic (PLEG): container finished" podID="1d13d0a6-ede9-45fe-9338-5a90260f0455" containerID="fd596c2c4a357feb978a9acf180d8d3ae8340b1b043c17107ff33fe257d0ce7d" exitCode=0 Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.407343 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1d13d0a6-ede9-45fe-9338-5a90260f0455","Type":"ContainerDied","Data":"e6cf52ef3f84a1b3eede7e7564cb23f48e1ed825f07acb54857b63501d13c82a"} Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.407383 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1d13d0a6-ede9-45fe-9338-5a90260f0455","Type":"ContainerDied","Data":"3acc162dd29cb480c5e3a7b72d533b66a8a7fa5367914e43729749c74ad05402"} Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.407395 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1d13d0a6-ede9-45fe-9338-5a90260f0455","Type":"ContainerDied","Data":"fd596c2c4a357feb978a9acf180d8d3ae8340b1b043c17107ff33fe257d0ce7d"} Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.434807 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpdgk\" (UniqueName: \"kubernetes.io/projected/0f0f40ab-1168-4fea-84b1-b134eb2a1b3c-kube-api-access-zpdgk\") pod \"0f0f40ab-1168-4fea-84b1-b134eb2a1b3c\" (UID: \"0f0f40ab-1168-4fea-84b1-b134eb2a1b3c\") " Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.434901 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f0f40ab-1168-4fea-84b1-b134eb2a1b3c-combined-ca-bundle\") pod \"0f0f40ab-1168-4fea-84b1-b134eb2a1b3c\" (UID: \"0f0f40ab-1168-4fea-84b1-b134eb2a1b3c\") " Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.434965 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f0f40ab-1168-4fea-84b1-b134eb2a1b3c-config-data\") pod \"0f0f40ab-1168-4fea-84b1-b134eb2a1b3c\" (UID: \"0f0f40ab-1168-4fea-84b1-b134eb2a1b3c\") " Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.439547 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f0f40ab-1168-4fea-84b1-b134eb2a1b3c-kube-api-access-zpdgk" (OuterVolumeSpecName: "kube-api-access-zpdgk") pod "0f0f40ab-1168-4fea-84b1-b134eb2a1b3c" (UID: "0f0f40ab-1168-4fea-84b1-b134eb2a1b3c"). InnerVolumeSpecName "kube-api-access-zpdgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.468340 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f0f40ab-1168-4fea-84b1-b134eb2a1b3c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f0f40ab-1168-4fea-84b1-b134eb2a1b3c" (UID: "0f0f40ab-1168-4fea-84b1-b134eb2a1b3c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.478097 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f0f40ab-1168-4fea-84b1-b134eb2a1b3c-config-data" (OuterVolumeSpecName: "config-data") pod "0f0f40ab-1168-4fea-84b1-b134eb2a1b3c" (UID: "0f0f40ab-1168-4fea-84b1-b134eb2a1b3c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.539026 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpdgk\" (UniqueName: \"kubernetes.io/projected/0f0f40ab-1168-4fea-84b1-b134eb2a1b3c-kube-api-access-zpdgk\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.539111 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f0f40ab-1168-4fea-84b1-b134eb2a1b3c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.539136 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f0f40ab-1168-4fea-84b1-b134eb2a1b3c-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.638554 4934 scope.go:117] "RemoveContainer" containerID="bc6848ec46001daa30195a2f745ad7629c7bf68dd56b59d8b02fda09322ce597" Dec 27 08:07:13 crc kubenswrapper[4934]: E1227 08:07:13.638966 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc6848ec46001daa30195a2f745ad7629c7bf68dd56b59d8b02fda09322ce597\": container with ID starting with bc6848ec46001daa30195a2f745ad7629c7bf68dd56b59d8b02fda09322ce597 not found: ID does not exist" containerID="bc6848ec46001daa30195a2f745ad7629c7bf68dd56b59d8b02fda09322ce597" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.638999 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc6848ec46001daa30195a2f745ad7629c7bf68dd56b59d8b02fda09322ce597"} err="failed to get container status \"bc6848ec46001daa30195a2f745ad7629c7bf68dd56b59d8b02fda09322ce597\": rpc error: code = NotFound desc = could not find container \"bc6848ec46001daa30195a2f745ad7629c7bf68dd56b59d8b02fda09322ce597\": container with ID starting with bc6848ec46001daa30195a2f745ad7629c7bf68dd56b59d8b02fda09322ce597 not found: ID does not exist" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.654023 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.714179 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.734118 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.742712 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thlkl\" (UniqueName: \"kubernetes.io/projected/4c696b92-7128-4465-875f-d2ef4298e798-kube-api-access-thlkl\") pod \"4c696b92-7128-4465-875f-d2ef4298e798\" (UID: \"4c696b92-7128-4465-875f-d2ef4298e798\") " Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.742833 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c696b92-7128-4465-875f-d2ef4298e798-combined-ca-bundle\") pod \"4c696b92-7128-4465-875f-d2ef4298e798\" (UID: \"4c696b92-7128-4465-875f-d2ef4298e798\") " Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.743057 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c696b92-7128-4465-875f-d2ef4298e798-config-data\") pod \"4c696b92-7128-4465-875f-d2ef4298e798\" (UID: \"4c696b92-7128-4465-875f-d2ef4298e798\") " Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.743182 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c696b92-7128-4465-875f-d2ef4298e798-logs\") pod \"4c696b92-7128-4465-875f-d2ef4298e798\" (UID: \"4c696b92-7128-4465-875f-d2ef4298e798\") " Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.744627 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c696b92-7128-4465-875f-d2ef4298e798-logs" (OuterVolumeSpecName: "logs") pod "4c696b92-7128-4465-875f-d2ef4298e798" (UID: "4c696b92-7128-4465-875f-d2ef4298e798"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.749237 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 27 08:07:13 crc kubenswrapper[4934]: E1227 08:07:13.749950 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c696b92-7128-4465-875f-d2ef4298e798" containerName="nova-metadata-metadata" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.749973 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c696b92-7128-4465-875f-d2ef4298e798" containerName="nova-metadata-metadata" Dec 27 08:07:13 crc kubenswrapper[4934]: E1227 08:07:13.750006 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c696b92-7128-4465-875f-d2ef4298e798" containerName="nova-metadata-log" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.750015 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c696b92-7128-4465-875f-d2ef4298e798" containerName="nova-metadata-log" Dec 27 08:07:13 crc kubenswrapper[4934]: E1227 08:07:13.750034 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f0f40ab-1168-4fea-84b1-b134eb2a1b3c" containerName="nova-cell1-novncproxy-novncproxy" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.750044 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f0f40ab-1168-4fea-84b1-b134eb2a1b3c" containerName="nova-cell1-novncproxy-novncproxy" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.750499 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f0f40ab-1168-4fea-84b1-b134eb2a1b3c" containerName="nova-cell1-novncproxy-novncproxy" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.750570 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c696b92-7128-4465-875f-d2ef4298e798" containerName="nova-metadata-metadata" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.750597 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c696b92-7128-4465-875f-d2ef4298e798" containerName="nova-metadata-log" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.751867 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.754094 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c696b92-7128-4465-875f-d2ef4298e798-kube-api-access-thlkl" (OuterVolumeSpecName: "kube-api-access-thlkl") pod "4c696b92-7128-4465-875f-d2ef4298e798" (UID: "4c696b92-7128-4465-875f-d2ef4298e798"). InnerVolumeSpecName "kube-api-access-thlkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.758067 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.758499 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.759009 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.759134 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.796300 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c696b92-7128-4465-875f-d2ef4298e798-config-data" (OuterVolumeSpecName: "config-data") pod "4c696b92-7128-4465-875f-d2ef4298e798" (UID: "4c696b92-7128-4465-875f-d2ef4298e798"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.799919 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c696b92-7128-4465-875f-d2ef4298e798-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c696b92-7128-4465-875f-d2ef4298e798" (UID: "4c696b92-7128-4465-875f-d2ef4298e798"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.846492 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be9c97e5-5a63-44cd-854f-16a1bcbbeb35-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"be9c97e5-5a63-44cd-854f-16a1bcbbeb35\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.846734 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/be9c97e5-5a63-44cd-854f-16a1bcbbeb35-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"be9c97e5-5a63-44cd-854f-16a1bcbbeb35\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.846807 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/be9c97e5-5a63-44cd-854f-16a1bcbbeb35-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"be9c97e5-5a63-44cd-854f-16a1bcbbeb35\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.847148 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be9c97e5-5a63-44cd-854f-16a1bcbbeb35-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"be9c97e5-5a63-44cd-854f-16a1bcbbeb35\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.847456 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br5tc\" (UniqueName: \"kubernetes.io/projected/be9c97e5-5a63-44cd-854f-16a1bcbbeb35-kube-api-access-br5tc\") pod \"nova-cell1-novncproxy-0\" (UID: \"be9c97e5-5a63-44cd-854f-16a1bcbbeb35\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.847651 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c696b92-7128-4465-875f-d2ef4298e798-logs\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.847668 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thlkl\" (UniqueName: \"kubernetes.io/projected/4c696b92-7128-4465-875f-d2ef4298e798-kube-api-access-thlkl\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.847680 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c696b92-7128-4465-875f-d2ef4298e798-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.847690 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c696b92-7128-4465-875f-d2ef4298e798-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.949803 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/be9c97e5-5a63-44cd-854f-16a1bcbbeb35-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"be9c97e5-5a63-44cd-854f-16a1bcbbeb35\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.950129 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/be9c97e5-5a63-44cd-854f-16a1bcbbeb35-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"be9c97e5-5a63-44cd-854f-16a1bcbbeb35\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.950236 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be9c97e5-5a63-44cd-854f-16a1bcbbeb35-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"be9c97e5-5a63-44cd-854f-16a1bcbbeb35\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.950293 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br5tc\" (UniqueName: \"kubernetes.io/projected/be9c97e5-5a63-44cd-854f-16a1bcbbeb35-kube-api-access-br5tc\") pod \"nova-cell1-novncproxy-0\" (UID: \"be9c97e5-5a63-44cd-854f-16a1bcbbeb35\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.950317 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be9c97e5-5a63-44cd-854f-16a1bcbbeb35-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"be9c97e5-5a63-44cd-854f-16a1bcbbeb35\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.954865 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be9c97e5-5a63-44cd-854f-16a1bcbbeb35-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"be9c97e5-5a63-44cd-854f-16a1bcbbeb35\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.956710 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/be9c97e5-5a63-44cd-854f-16a1bcbbeb35-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"be9c97e5-5a63-44cd-854f-16a1bcbbeb35\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.956916 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/be9c97e5-5a63-44cd-854f-16a1bcbbeb35-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"be9c97e5-5a63-44cd-854f-16a1bcbbeb35\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.957257 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be9c97e5-5a63-44cd-854f-16a1bcbbeb35-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"be9c97e5-5a63-44cd-854f-16a1bcbbeb35\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:13 crc kubenswrapper[4934]: I1227 08:07:13.968549 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br5tc\" (UniqueName: \"kubernetes.io/projected/be9c97e5-5a63-44cd-854f-16a1bcbbeb35-kube-api-access-br5tc\") pod \"nova-cell1-novncproxy-0\" (UID: \"be9c97e5-5a63-44cd-854f-16a1bcbbeb35\") " pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.083392 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.421956 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2fad2302-c4ec-4c95-a25e-e0ef03a99b61","Type":"ContainerStarted","Data":"a838d1fc6cda2e74d9bbbcaa37c34ec07dd0d7043fc02ddaa29abb61ab58ab7e"} Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.426423 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4c696b92-7128-4465-875f-d2ef4298e798","Type":"ContainerDied","Data":"1a78506bb3f041e1daf078de7f37bad24b4eca3082f9942981c516c2990b591f"} Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.426460 4934 scope.go:117] "RemoveContainer" containerID="bc9b8337126f66220fa9bcfe2a59c8bf4423fcc2292e6bbc9bf9ca725a983c0b" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.426632 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.449285 4934 scope.go:117] "RemoveContainer" containerID="0e363c5e450ac1ceba74fb25171c16da42c41caad9e6478a642ac132a0d9e0af" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.478169 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.493812 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.507727 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.509806 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.511608 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.512051 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.520027 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.567471 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.672729 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97621094-4d07-4ed0-bf11-82bf65dde763-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"97621094-4d07-4ed0-bf11-82bf65dde763\") " pod="openstack/nova-metadata-0" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.673121 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97621094-4d07-4ed0-bf11-82bf65dde763-logs\") pod \"nova-metadata-0\" (UID: \"97621094-4d07-4ed0-bf11-82bf65dde763\") " pod="openstack/nova-metadata-0" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.673318 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97621094-4d07-4ed0-bf11-82bf65dde763-config-data\") pod \"nova-metadata-0\" (UID: \"97621094-4d07-4ed0-bf11-82bf65dde763\") " pod="openstack/nova-metadata-0" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.673505 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl6ln\" (UniqueName: \"kubernetes.io/projected/97621094-4d07-4ed0-bf11-82bf65dde763-kube-api-access-rl6ln\") pod \"nova-metadata-0\" (UID: \"97621094-4d07-4ed0-bf11-82bf65dde763\") " pod="openstack/nova-metadata-0" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.673791 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/97621094-4d07-4ed0-bf11-82bf65dde763-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"97621094-4d07-4ed0-bf11-82bf65dde763\") " pod="openstack/nova-metadata-0" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.777310 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97621094-4d07-4ed0-bf11-82bf65dde763-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"97621094-4d07-4ed0-bf11-82bf65dde763\") " pod="openstack/nova-metadata-0" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.777361 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97621094-4d07-4ed0-bf11-82bf65dde763-logs\") pod \"nova-metadata-0\" (UID: \"97621094-4d07-4ed0-bf11-82bf65dde763\") " pod="openstack/nova-metadata-0" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.777387 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97621094-4d07-4ed0-bf11-82bf65dde763-config-data\") pod \"nova-metadata-0\" (UID: \"97621094-4d07-4ed0-bf11-82bf65dde763\") " pod="openstack/nova-metadata-0" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.777409 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl6ln\" (UniqueName: \"kubernetes.io/projected/97621094-4d07-4ed0-bf11-82bf65dde763-kube-api-access-rl6ln\") pod \"nova-metadata-0\" (UID: \"97621094-4d07-4ed0-bf11-82bf65dde763\") " pod="openstack/nova-metadata-0" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.777479 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/97621094-4d07-4ed0-bf11-82bf65dde763-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"97621094-4d07-4ed0-bf11-82bf65dde763\") " pod="openstack/nova-metadata-0" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.778323 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97621094-4d07-4ed0-bf11-82bf65dde763-logs\") pod \"nova-metadata-0\" (UID: \"97621094-4d07-4ed0-bf11-82bf65dde763\") " pod="openstack/nova-metadata-0" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.782118 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97621094-4d07-4ed0-bf11-82bf65dde763-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"97621094-4d07-4ed0-bf11-82bf65dde763\") " pod="openstack/nova-metadata-0" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.784307 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/97621094-4d07-4ed0-bf11-82bf65dde763-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"97621094-4d07-4ed0-bf11-82bf65dde763\") " pod="openstack/nova-metadata-0" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.786263 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97621094-4d07-4ed0-bf11-82bf65dde763-config-data\") pod \"nova-metadata-0\" (UID: \"97621094-4d07-4ed0-bf11-82bf65dde763\") " pod="openstack/nova-metadata-0" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.799359 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl6ln\" (UniqueName: \"kubernetes.io/projected/97621094-4d07-4ed0-bf11-82bf65dde763-kube-api-access-rl6ln\") pod \"nova-metadata-0\" (UID: \"97621094-4d07-4ed0-bf11-82bf65dde763\") " pod="openstack/nova-metadata-0" Dec 27 08:07:14 crc kubenswrapper[4934]: I1227 08:07:14.827976 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 27 08:07:15 crc kubenswrapper[4934]: I1227 08:07:15.330587 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:07:15 crc kubenswrapper[4934]: I1227 08:07:15.331070 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:07:15 crc kubenswrapper[4934]: I1227 08:07:15.340426 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 27 08:07:15 crc kubenswrapper[4934]: W1227 08:07:15.341352 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97621094_4d07_4ed0_bf11_82bf65dde763.slice/crio-c8a4abf616f5e08edc844bf7a42aacbb2f87fbe013e378d98d4f07cea658345b WatchSource:0}: Error finding container c8a4abf616f5e08edc844bf7a42aacbb2f87fbe013e378d98d4f07cea658345b: Status 404 returned error can't find the container with id c8a4abf616f5e08edc844bf7a42aacbb2f87fbe013e378d98d4f07cea658345b Dec 27 08:07:15 crc kubenswrapper[4934]: I1227 08:07:15.441412 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2fad2302-c4ec-4c95-a25e-e0ef03a99b61","Type":"ContainerStarted","Data":"173a5896a39835fe558ed789ac2e7bf5099dbbf22c767f3a4e27cb1fc8cbb196"} Dec 27 08:07:15 crc kubenswrapper[4934]: I1227 08:07:15.441619 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2fad2302-c4ec-4c95-a25e-e0ef03a99b61" containerName="ceilometer-central-agent" containerID="cri-o://3fa1c7010492f1fe94460776a4a760f0d0e68513d6db66ec3b278d16f660996f" gracePeriod=30 Dec 27 08:07:15 crc kubenswrapper[4934]: I1227 08:07:15.441897 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 27 08:07:15 crc kubenswrapper[4934]: I1227 08:07:15.442342 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2fad2302-c4ec-4c95-a25e-e0ef03a99b61" containerName="proxy-httpd" containerID="cri-o://173a5896a39835fe558ed789ac2e7bf5099dbbf22c767f3a4e27cb1fc8cbb196" gracePeriod=30 Dec 27 08:07:15 crc kubenswrapper[4934]: I1227 08:07:15.442404 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2fad2302-c4ec-4c95-a25e-e0ef03a99b61" containerName="ceilometer-notification-agent" containerID="cri-o://a1159f38cea2415aafb205147cca11c7cff559a38e5b1821fbfe82ff2e7fee1f" gracePeriod=30 Dec 27 08:07:15 crc kubenswrapper[4934]: I1227 08:07:15.442425 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2fad2302-c4ec-4c95-a25e-e0ef03a99b61" containerName="sg-core" containerID="cri-o://a838d1fc6cda2e74d9bbbcaa37c34ec07dd0d7043fc02ddaa29abb61ab58ab7e" gracePeriod=30 Dec 27 08:07:15 crc kubenswrapper[4934]: I1227 08:07:15.453648 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"be9c97e5-5a63-44cd-854f-16a1bcbbeb35","Type":"ContainerStarted","Data":"54761fe49609e8b841e5ab1cc93a9c0fac2c4d1a938b897dbae9ffa1ab8e40a6"} Dec 27 08:07:15 crc kubenswrapper[4934]: I1227 08:07:15.453703 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"be9c97e5-5a63-44cd-854f-16a1bcbbeb35","Type":"ContainerStarted","Data":"5d026d09edd31dd19486ffa76341e8da52fe4b3adc221edbde91130c21761430"} Dec 27 08:07:15 crc kubenswrapper[4934]: I1227 08:07:15.458195 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"97621094-4d07-4ed0-bf11-82bf65dde763","Type":"ContainerStarted","Data":"c8a4abf616f5e08edc844bf7a42aacbb2f87fbe013e378d98d4f07cea658345b"} Dec 27 08:07:15 crc kubenswrapper[4934]: I1227 08:07:15.465309 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.10608744 podStartE2EDuration="7.465291663s" podCreationTimestamp="2025-12-27 08:07:08 +0000 UTC" firstStartedPulling="2025-12-27 08:07:09.378707048 +0000 UTC m=+1490.199147642" lastFinishedPulling="2025-12-27 08:07:14.737911281 +0000 UTC m=+1495.558351865" observedRunningTime="2025-12-27 08:07:15.464288979 +0000 UTC m=+1496.284729583" watchObservedRunningTime="2025-12-27 08:07:15.465291663 +0000 UTC m=+1496.285732257" Dec 27 08:07:15 crc kubenswrapper[4934]: I1227 08:07:15.493193 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f0f40ab-1168-4fea-84b1-b134eb2a1b3c" path="/var/lib/kubelet/pods/0f0f40ab-1168-4fea-84b1-b134eb2a1b3c/volumes" Dec 27 08:07:15 crc kubenswrapper[4934]: I1227 08:07:15.494495 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c696b92-7128-4465-875f-d2ef4298e798" path="/var/lib/kubelet/pods/4c696b92-7128-4465-875f-d2ef4298e798/volumes" Dec 27 08:07:15 crc kubenswrapper[4934]: I1227 08:07:15.499793 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.499780359 podStartE2EDuration="2.499780359s" podCreationTimestamp="2025-12-27 08:07:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:07:15.488843725 +0000 UTC m=+1496.309284319" watchObservedRunningTime="2025-12-27 08:07:15.499780359 +0000 UTC m=+1496.320220953" Dec 27 08:07:15 crc kubenswrapper[4934]: I1227 08:07:15.567286 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 27 08:07:15 crc kubenswrapper[4934]: I1227 08:07:15.568002 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 27 08:07:15 crc kubenswrapper[4934]: I1227 08:07:15.572492 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 27 08:07:15 crc kubenswrapper[4934]: I1227 08:07:15.574770 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.483780 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"97621094-4d07-4ed0-bf11-82bf65dde763","Type":"ContainerStarted","Data":"b5af56b2b2efd0e89446980bdc37b9a8cd92dc2a5d1cde92aadbe4006664bde3"} Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.483840 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"97621094-4d07-4ed0-bf11-82bf65dde763","Type":"ContainerStarted","Data":"5305c4245d99ea632dccdb5b3b9f627bd571d2a06383351642303ebf8a8521b5"} Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.488709 4934 generic.go:334] "Generic (PLEG): container finished" podID="2fad2302-c4ec-4c95-a25e-e0ef03a99b61" containerID="173a5896a39835fe558ed789ac2e7bf5099dbbf22c767f3a4e27cb1fc8cbb196" exitCode=0 Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.488743 4934 generic.go:334] "Generic (PLEG): container finished" podID="2fad2302-c4ec-4c95-a25e-e0ef03a99b61" containerID="a838d1fc6cda2e74d9bbbcaa37c34ec07dd0d7043fc02ddaa29abb61ab58ab7e" exitCode=2 Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.488756 4934 generic.go:334] "Generic (PLEG): container finished" podID="2fad2302-c4ec-4c95-a25e-e0ef03a99b61" containerID="a1159f38cea2415aafb205147cca11c7cff559a38e5b1821fbfe82ff2e7fee1f" exitCode=0 Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.488995 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2fad2302-c4ec-4c95-a25e-e0ef03a99b61","Type":"ContainerDied","Data":"173a5896a39835fe558ed789ac2e7bf5099dbbf22c767f3a4e27cb1fc8cbb196"} Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.489035 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2fad2302-c4ec-4c95-a25e-e0ef03a99b61","Type":"ContainerDied","Data":"a838d1fc6cda2e74d9bbbcaa37c34ec07dd0d7043fc02ddaa29abb61ab58ab7e"} Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.489049 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2fad2302-c4ec-4c95-a25e-e0ef03a99b61","Type":"ContainerDied","Data":"a1159f38cea2415aafb205147cca11c7cff559a38e5b1821fbfe82ff2e7fee1f"} Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.489352 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.508818 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.518752 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.518727999 podStartE2EDuration="2.518727999s" podCreationTimestamp="2025-12-27 08:07:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:07:16.506445286 +0000 UTC m=+1497.326885900" watchObservedRunningTime="2025-12-27 08:07:16.518727999 +0000 UTC m=+1497.339168613" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.681607 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-mr67x"] Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.687402 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.692085 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-mr67x"] Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.833620 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-config\") pod \"dnsmasq-dns-f84f9ccf-mr67x\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.833956 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s7ct\" (UniqueName: \"kubernetes.io/projected/1cede442-232c-45b8-a8d3-7d617531d53c-kube-api-access-4s7ct\") pod \"dnsmasq-dns-f84f9ccf-mr67x\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.834263 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-ovsdbserver-sb\") pod \"dnsmasq-dns-f84f9ccf-mr67x\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.834314 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-ovsdbserver-nb\") pod \"dnsmasq-dns-f84f9ccf-mr67x\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.834510 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-dns-swift-storage-0\") pod \"dnsmasq-dns-f84f9ccf-mr67x\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.834720 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-dns-svc\") pod \"dnsmasq-dns-f84f9ccf-mr67x\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.936426 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s7ct\" (UniqueName: \"kubernetes.io/projected/1cede442-232c-45b8-a8d3-7d617531d53c-kube-api-access-4s7ct\") pod \"dnsmasq-dns-f84f9ccf-mr67x\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.936812 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-ovsdbserver-sb\") pod \"dnsmasq-dns-f84f9ccf-mr67x\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.936940 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-ovsdbserver-nb\") pod \"dnsmasq-dns-f84f9ccf-mr67x\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.937126 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-dns-swift-storage-0\") pod \"dnsmasq-dns-f84f9ccf-mr67x\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.937293 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-dns-svc\") pod \"dnsmasq-dns-f84f9ccf-mr67x\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.937457 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-config\") pod \"dnsmasq-dns-f84f9ccf-mr67x\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.937863 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-ovsdbserver-sb\") pod \"dnsmasq-dns-f84f9ccf-mr67x\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.937942 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-ovsdbserver-nb\") pod \"dnsmasq-dns-f84f9ccf-mr67x\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.937971 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-dns-swift-storage-0\") pod \"dnsmasq-dns-f84f9ccf-mr67x\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.938114 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-dns-svc\") pod \"dnsmasq-dns-f84f9ccf-mr67x\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.938257 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-config\") pod \"dnsmasq-dns-f84f9ccf-mr67x\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:16 crc kubenswrapper[4934]: I1227 08:07:16.967920 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s7ct\" (UniqueName: \"kubernetes.io/projected/1cede442-232c-45b8-a8d3-7d617531d53c-kube-api-access-4s7ct\") pod \"dnsmasq-dns-f84f9ccf-mr67x\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:17 crc kubenswrapper[4934]: I1227 08:07:17.019260 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:17 crc kubenswrapper[4934]: I1227 08:07:17.531083 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-mr67x"] Dec 27 08:07:17 crc kubenswrapper[4934]: I1227 08:07:17.552087 4934 generic.go:334] "Generic (PLEG): container finished" podID="2fad2302-c4ec-4c95-a25e-e0ef03a99b61" containerID="3fa1c7010492f1fe94460776a4a760f0d0e68513d6db66ec3b278d16f660996f" exitCode=0 Dec 27 08:07:17 crc kubenswrapper[4934]: I1227 08:07:17.553274 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2fad2302-c4ec-4c95-a25e-e0ef03a99b61","Type":"ContainerDied","Data":"3fa1c7010492f1fe94460776a4a760f0d0e68513d6db66ec3b278d16f660996f"} Dec 27 08:07:17 crc kubenswrapper[4934]: I1227 08:07:17.942410 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.071406 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-combined-ca-bundle\") pod \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.071495 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frmkn\" (UniqueName: \"kubernetes.io/projected/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-kube-api-access-frmkn\") pod \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.071551 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-log-httpd\") pod \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.071592 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-run-httpd\") pod \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.071684 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-sg-core-conf-yaml\") pod \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.071751 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-config-data\") pod \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.071825 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-scripts\") pod \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\" (UID: \"2fad2302-c4ec-4c95-a25e-e0ef03a99b61\") " Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.072631 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2fad2302-c4ec-4c95-a25e-e0ef03a99b61" (UID: "2fad2302-c4ec-4c95-a25e-e0ef03a99b61"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.073658 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2fad2302-c4ec-4c95-a25e-e0ef03a99b61" (UID: "2fad2302-c4ec-4c95-a25e-e0ef03a99b61"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.077844 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-scripts" (OuterVolumeSpecName: "scripts") pod "2fad2302-c4ec-4c95-a25e-e0ef03a99b61" (UID: "2fad2302-c4ec-4c95-a25e-e0ef03a99b61"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.077921 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-kube-api-access-frmkn" (OuterVolumeSpecName: "kube-api-access-frmkn") pod "2fad2302-c4ec-4c95-a25e-e0ef03a99b61" (UID: "2fad2302-c4ec-4c95-a25e-e0ef03a99b61"). InnerVolumeSpecName "kube-api-access-frmkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.124360 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2fad2302-c4ec-4c95-a25e-e0ef03a99b61" (UID: "2fad2302-c4ec-4c95-a25e-e0ef03a99b61"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.174969 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frmkn\" (UniqueName: \"kubernetes.io/projected/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-kube-api-access-frmkn\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.175005 4934 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.175019 4934 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.175030 4934 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.175039 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.182812 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2fad2302-c4ec-4c95-a25e-e0ef03a99b61" (UID: "2fad2302-c4ec-4c95-a25e-e0ef03a99b61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.248481 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-config-data" (OuterVolumeSpecName: "config-data") pod "2fad2302-c4ec-4c95-a25e-e0ef03a99b61" (UID: "2fad2302-c4ec-4c95-a25e-e0ef03a99b61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.276999 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.277030 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fad2302-c4ec-4c95-a25e-e0ef03a99b61-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.564010 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.563997 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2fad2302-c4ec-4c95-a25e-e0ef03a99b61","Type":"ContainerDied","Data":"75ebb1cbce396521261708603d5b5be37fc812e042adf87066ab5b10f460be16"} Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.564129 4934 scope.go:117] "RemoveContainer" containerID="173a5896a39835fe558ed789ac2e7bf5099dbbf22c767f3a4e27cb1fc8cbb196" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.565897 4934 generic.go:334] "Generic (PLEG): container finished" podID="1cede442-232c-45b8-a8d3-7d617531d53c" containerID="b5ea68957cea70f159ffd8d767539802def518bcb94ba522afc2a1da848e2497" exitCode=0 Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.565985 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" event={"ID":"1cede442-232c-45b8-a8d3-7d617531d53c","Type":"ContainerDied","Data":"b5ea68957cea70f159ffd8d767539802def518bcb94ba522afc2a1da848e2497"} Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.566050 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" event={"ID":"1cede442-232c-45b8-a8d3-7d617531d53c","Type":"ContainerStarted","Data":"2ee369df8613cf716d77563c55db97d2550da30e720365dbdbd0e2852227d276"} Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.603769 4934 scope.go:117] "RemoveContainer" containerID="a838d1fc6cda2e74d9bbbcaa37c34ec07dd0d7043fc02ddaa29abb61ab58ab7e" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.768781 4934 scope.go:117] "RemoveContainer" containerID="a1159f38cea2415aafb205147cca11c7cff559a38e5b1821fbfe82ff2e7fee1f" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.792607 4934 scope.go:117] "RemoveContainer" containerID="3fa1c7010492f1fe94460776a4a760f0d0e68513d6db66ec3b278d16f660996f" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.800626 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.812199 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.821903 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:07:18 crc kubenswrapper[4934]: E1227 08:07:18.822583 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fad2302-c4ec-4c95-a25e-e0ef03a99b61" containerName="sg-core" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.822602 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fad2302-c4ec-4c95-a25e-e0ef03a99b61" containerName="sg-core" Dec 27 08:07:18 crc kubenswrapper[4934]: E1227 08:07:18.822625 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fad2302-c4ec-4c95-a25e-e0ef03a99b61" containerName="ceilometer-notification-agent" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.822631 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fad2302-c4ec-4c95-a25e-e0ef03a99b61" containerName="ceilometer-notification-agent" Dec 27 08:07:18 crc kubenswrapper[4934]: E1227 08:07:18.822644 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fad2302-c4ec-4c95-a25e-e0ef03a99b61" containerName="proxy-httpd" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.822650 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fad2302-c4ec-4c95-a25e-e0ef03a99b61" containerName="proxy-httpd" Dec 27 08:07:18 crc kubenswrapper[4934]: E1227 08:07:18.822676 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fad2302-c4ec-4c95-a25e-e0ef03a99b61" containerName="ceilometer-central-agent" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.822682 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fad2302-c4ec-4c95-a25e-e0ef03a99b61" containerName="ceilometer-central-agent" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.822888 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fad2302-c4ec-4c95-a25e-e0ef03a99b61" containerName="sg-core" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.822921 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fad2302-c4ec-4c95-a25e-e0ef03a99b61" containerName="proxy-httpd" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.822938 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fad2302-c4ec-4c95-a25e-e0ef03a99b61" containerName="ceilometer-central-agent" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.822959 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fad2302-c4ec-4c95-a25e-e0ef03a99b61" containerName="ceilometer-notification-agent" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.828144 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.830199 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.830381 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.836177 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.991394 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " pod="openstack/ceilometer-0" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.991671 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgthd\" (UniqueName: \"kubernetes.io/projected/8c109ea6-23b9-426e-a44b-44592c85b112-kube-api-access-jgthd\") pod \"ceilometer-0\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " pod="openstack/ceilometer-0" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.991705 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " pod="openstack/ceilometer-0" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.991894 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-scripts\") pod \"ceilometer-0\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " pod="openstack/ceilometer-0" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.991972 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c109ea6-23b9-426e-a44b-44592c85b112-log-httpd\") pod \"ceilometer-0\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " pod="openstack/ceilometer-0" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.992085 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c109ea6-23b9-426e-a44b-44592c85b112-run-httpd\") pod \"ceilometer-0\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " pod="openstack/ceilometer-0" Dec 27 08:07:18 crc kubenswrapper[4934]: I1227 08:07:18.992192 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-config-data\") pod \"ceilometer-0\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " pod="openstack/ceilometer-0" Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.084367 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.093708 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c109ea6-23b9-426e-a44b-44592c85b112-log-httpd\") pod \"ceilometer-0\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " pod="openstack/ceilometer-0" Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.093775 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c109ea6-23b9-426e-a44b-44592c85b112-run-httpd\") pod \"ceilometer-0\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " pod="openstack/ceilometer-0" Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.093823 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-config-data\") pod \"ceilometer-0\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " pod="openstack/ceilometer-0" Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.093899 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " pod="openstack/ceilometer-0" Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.093945 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgthd\" (UniqueName: \"kubernetes.io/projected/8c109ea6-23b9-426e-a44b-44592c85b112-kube-api-access-jgthd\") pod \"ceilometer-0\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " pod="openstack/ceilometer-0" Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.093978 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " pod="openstack/ceilometer-0" Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.094060 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-scripts\") pod \"ceilometer-0\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " pod="openstack/ceilometer-0" Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.094225 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c109ea6-23b9-426e-a44b-44592c85b112-log-httpd\") pod \"ceilometer-0\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " pod="openstack/ceilometer-0" Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.094682 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c109ea6-23b9-426e-a44b-44592c85b112-run-httpd\") pod \"ceilometer-0\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " pod="openstack/ceilometer-0" Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.098581 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " pod="openstack/ceilometer-0" Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.099026 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-config-data\") pod \"ceilometer-0\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " pod="openstack/ceilometer-0" Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.099125 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " pod="openstack/ceilometer-0" Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.101285 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-scripts\") pod \"ceilometer-0\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " pod="openstack/ceilometer-0" Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.109872 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgthd\" (UniqueName: \"kubernetes.io/projected/8c109ea6-23b9-426e-a44b-44592c85b112-kube-api-access-jgthd\") pod \"ceilometer-0\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " pod="openstack/ceilometer-0" Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.150433 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.178339 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.329956 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.540328 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fad2302-c4ec-4c95-a25e-e0ef03a99b61" path="/var/lib/kubelet/pods/2fad2302-c4ec-4c95-a25e-e0ef03a99b61/volumes" Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.578803 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" event={"ID":"1cede442-232c-45b8-a8d3-7d617531d53c","Type":"ContainerStarted","Data":"012b79d6af6ea82699a9fade6fb224056c3ba52ab3a89131e4db688efd4fca9e"} Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.579255 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.581009 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9753e530-5b55-4951-af42-71297ce5c944" containerName="nova-api-log" containerID="cri-o://f8f94b18eff1c88a06c4d34a79292ae09139ccfbc7cd8e5fae2b49b1571e74e8" gracePeriod=30 Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.581237 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9753e530-5b55-4951-af42-71297ce5c944" containerName="nova-api-api" containerID="cri-o://a049ffa669f4d9521ebf64ba2c09b7a1ac770f1963ed43a509e333623eca0f4e" gracePeriod=30 Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.607791 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" podStartSLOduration=3.607774046 podStartE2EDuration="3.607774046s" podCreationTimestamp="2025-12-27 08:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:07:19.597670703 +0000 UTC m=+1500.418111297" watchObservedRunningTime="2025-12-27 08:07:19.607774046 +0000 UTC m=+1500.428214640" Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.688643 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.828226 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 27 08:07:19 crc kubenswrapper[4934]: I1227 08:07:19.828960 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 27 08:07:20 crc kubenswrapper[4934]: I1227 08:07:20.605922 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c109ea6-23b9-426e-a44b-44592c85b112","Type":"ContainerStarted","Data":"d9dc27775e1b7438b8462a95b980ea4be4d83f6ae3a8218900ccbfcfeee20ee6"} Dec 27 08:07:20 crc kubenswrapper[4934]: I1227 08:07:20.608195 4934 generic.go:334] "Generic (PLEG): container finished" podID="9753e530-5b55-4951-af42-71297ce5c944" containerID="f8f94b18eff1c88a06c4d34a79292ae09139ccfbc7cd8e5fae2b49b1571e74e8" exitCode=143 Dec 27 08:07:20 crc kubenswrapper[4934]: I1227 08:07:20.609281 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9753e530-5b55-4951-af42-71297ce5c944","Type":"ContainerDied","Data":"f8f94b18eff1c88a06c4d34a79292ae09139ccfbc7cd8e5fae2b49b1571e74e8"} Dec 27 08:07:21 crc kubenswrapper[4934]: I1227 08:07:21.621660 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c109ea6-23b9-426e-a44b-44592c85b112","Type":"ContainerStarted","Data":"f90732ae6832c4b43f77d0cb5f281af5d28022f942eb806736fda794504c2ea0"} Dec 27 08:07:21 crc kubenswrapper[4934]: I1227 08:07:21.622238 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c109ea6-23b9-426e-a44b-44592c85b112","Type":"ContainerStarted","Data":"f413a4ac455122d4bf53aee391af087cad4883b8bc0283d8b97ff272f43b2bcb"} Dec 27 08:07:22 crc kubenswrapper[4934]: I1227 08:07:22.643180 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c109ea6-23b9-426e-a44b-44592c85b112","Type":"ContainerStarted","Data":"cd86a75c8fcf7377d8218e9ec6470b11b88cf7f8fa277455d763e25e1421efc6"} Dec 27 08:07:23 crc kubenswrapper[4934]: E1227 08:07:23.009117 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9753e530_5b55_4951_af42_71297ce5c944.slice/crio-f8f94b18eff1c88a06c4d34a79292ae09139ccfbc7cd8e5fae2b49b1571e74e8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9753e530_5b55_4951_af42_71297ce5c944.slice/crio-a049ffa669f4d9521ebf64ba2c09b7a1ac770f1963ed43a509e333623eca0f4e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9753e530_5b55_4951_af42_71297ce5c944.slice/crio-conmon-a049ffa669f4d9521ebf64ba2c09b7a1ac770f1963ed43a509e333623eca0f4e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9753e530_5b55_4951_af42_71297ce5c944.slice/crio-conmon-f8f94b18eff1c88a06c4d34a79292ae09139ccfbc7cd8e5fae2b49b1571e74e8.scope\": RecentStats: unable to find data in memory cache]" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.346663 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.423998 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jg5jp\" (UniqueName: \"kubernetes.io/projected/9753e530-5b55-4951-af42-71297ce5c944-kube-api-access-jg5jp\") pod \"9753e530-5b55-4951-af42-71297ce5c944\" (UID: \"9753e530-5b55-4951-af42-71297ce5c944\") " Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.424105 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9753e530-5b55-4951-af42-71297ce5c944-combined-ca-bundle\") pod \"9753e530-5b55-4951-af42-71297ce5c944\" (UID: \"9753e530-5b55-4951-af42-71297ce5c944\") " Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.424156 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9753e530-5b55-4951-af42-71297ce5c944-logs\") pod \"9753e530-5b55-4951-af42-71297ce5c944\" (UID: \"9753e530-5b55-4951-af42-71297ce5c944\") " Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.424303 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9753e530-5b55-4951-af42-71297ce5c944-config-data\") pod \"9753e530-5b55-4951-af42-71297ce5c944\" (UID: \"9753e530-5b55-4951-af42-71297ce5c944\") " Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.425435 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9753e530-5b55-4951-af42-71297ce5c944-logs" (OuterVolumeSpecName: "logs") pod "9753e530-5b55-4951-af42-71297ce5c944" (UID: "9753e530-5b55-4951-af42-71297ce5c944"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.434278 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9753e530-5b55-4951-af42-71297ce5c944-kube-api-access-jg5jp" (OuterVolumeSpecName: "kube-api-access-jg5jp") pod "9753e530-5b55-4951-af42-71297ce5c944" (UID: "9753e530-5b55-4951-af42-71297ce5c944"). InnerVolumeSpecName "kube-api-access-jg5jp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.465986 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9753e530-5b55-4951-af42-71297ce5c944-config-data" (OuterVolumeSpecName: "config-data") pod "9753e530-5b55-4951-af42-71297ce5c944" (UID: "9753e530-5b55-4951-af42-71297ce5c944"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.483845 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9753e530-5b55-4951-af42-71297ce5c944-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9753e530-5b55-4951-af42-71297ce5c944" (UID: "9753e530-5b55-4951-af42-71297ce5c944"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.527135 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9753e530-5b55-4951-af42-71297ce5c944-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.527696 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jg5jp\" (UniqueName: \"kubernetes.io/projected/9753e530-5b55-4951-af42-71297ce5c944-kube-api-access-jg5jp\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.527720 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9753e530-5b55-4951-af42-71297ce5c944-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.527733 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9753e530-5b55-4951-af42-71297ce5c944-logs\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.656713 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c109ea6-23b9-426e-a44b-44592c85b112","Type":"ContainerStarted","Data":"d197d9cc278fca8cccff864e6cea6401f9a3bcf90c29b139b1f86282c5aeede6"} Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.656954 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c109ea6-23b9-426e-a44b-44592c85b112" containerName="ceilometer-central-agent" containerID="cri-o://f413a4ac455122d4bf53aee391af087cad4883b8bc0283d8b97ff272f43b2bcb" gracePeriod=30 Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.657140 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.657700 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c109ea6-23b9-426e-a44b-44592c85b112" containerName="proxy-httpd" containerID="cri-o://d197d9cc278fca8cccff864e6cea6401f9a3bcf90c29b139b1f86282c5aeede6" gracePeriod=30 Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.657785 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c109ea6-23b9-426e-a44b-44592c85b112" containerName="sg-core" containerID="cri-o://cd86a75c8fcf7377d8218e9ec6470b11b88cf7f8fa277455d763e25e1421efc6" gracePeriod=30 Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.657880 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c109ea6-23b9-426e-a44b-44592c85b112" containerName="ceilometer-notification-agent" containerID="cri-o://f90732ae6832c4b43f77d0cb5f281af5d28022f942eb806736fda794504c2ea0" gracePeriod=30 Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.670849 4934 generic.go:334] "Generic (PLEG): container finished" podID="9753e530-5b55-4951-af42-71297ce5c944" containerID="a049ffa669f4d9521ebf64ba2c09b7a1ac770f1963ed43a509e333623eca0f4e" exitCode=0 Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.670898 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9753e530-5b55-4951-af42-71297ce5c944","Type":"ContainerDied","Data":"a049ffa669f4d9521ebf64ba2c09b7a1ac770f1963ed43a509e333623eca0f4e"} Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.670926 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9753e530-5b55-4951-af42-71297ce5c944","Type":"ContainerDied","Data":"4e6425a3dd6c65706918495dc5d63673b54f7119128a75c36fe95438cbcbd365"} Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.670952 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.670959 4934 scope.go:117] "RemoveContainer" containerID="a049ffa669f4d9521ebf64ba2c09b7a1ac770f1963ed43a509e333623eca0f4e" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.695486 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.220251167 podStartE2EDuration="5.695468595s" podCreationTimestamp="2025-12-27 08:07:18 +0000 UTC" firstStartedPulling="2025-12-27 08:07:19.698525065 +0000 UTC m=+1500.518965659" lastFinishedPulling="2025-12-27 08:07:23.173742483 +0000 UTC m=+1503.994183087" observedRunningTime="2025-12-27 08:07:23.681037612 +0000 UTC m=+1504.501478196" watchObservedRunningTime="2025-12-27 08:07:23.695468595 +0000 UTC m=+1504.515909189" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.702706 4934 scope.go:117] "RemoveContainer" containerID="f8f94b18eff1c88a06c4d34a79292ae09139ccfbc7cd8e5fae2b49b1571e74e8" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.719900 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.734502 4934 scope.go:117] "RemoveContainer" containerID="a049ffa669f4d9521ebf64ba2c09b7a1ac770f1963ed43a509e333623eca0f4e" Dec 27 08:07:23 crc kubenswrapper[4934]: E1227 08:07:23.736400 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a049ffa669f4d9521ebf64ba2c09b7a1ac770f1963ed43a509e333623eca0f4e\": container with ID starting with a049ffa669f4d9521ebf64ba2c09b7a1ac770f1963ed43a509e333623eca0f4e not found: ID does not exist" containerID="a049ffa669f4d9521ebf64ba2c09b7a1ac770f1963ed43a509e333623eca0f4e" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.736516 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a049ffa669f4d9521ebf64ba2c09b7a1ac770f1963ed43a509e333623eca0f4e"} err="failed to get container status \"a049ffa669f4d9521ebf64ba2c09b7a1ac770f1963ed43a509e333623eca0f4e\": rpc error: code = NotFound desc = could not find container \"a049ffa669f4d9521ebf64ba2c09b7a1ac770f1963ed43a509e333623eca0f4e\": container with ID starting with a049ffa669f4d9521ebf64ba2c09b7a1ac770f1963ed43a509e333623eca0f4e not found: ID does not exist" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.736620 4934 scope.go:117] "RemoveContainer" containerID="f8f94b18eff1c88a06c4d34a79292ae09139ccfbc7cd8e5fae2b49b1571e74e8" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.738610 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 27 08:07:23 crc kubenswrapper[4934]: E1227 08:07:23.738787 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8f94b18eff1c88a06c4d34a79292ae09139ccfbc7cd8e5fae2b49b1571e74e8\": container with ID starting with f8f94b18eff1c88a06c4d34a79292ae09139ccfbc7cd8e5fae2b49b1571e74e8 not found: ID does not exist" containerID="f8f94b18eff1c88a06c4d34a79292ae09139ccfbc7cd8e5fae2b49b1571e74e8" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.738885 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8f94b18eff1c88a06c4d34a79292ae09139ccfbc7cd8e5fae2b49b1571e74e8"} err="failed to get container status \"f8f94b18eff1c88a06c4d34a79292ae09139ccfbc7cd8e5fae2b49b1571e74e8\": rpc error: code = NotFound desc = could not find container \"f8f94b18eff1c88a06c4d34a79292ae09139ccfbc7cd8e5fae2b49b1571e74e8\": container with ID starting with f8f94b18eff1c88a06c4d34a79292ae09139ccfbc7cd8e5fae2b49b1571e74e8 not found: ID does not exist" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.767137 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 27 08:07:23 crc kubenswrapper[4934]: E1227 08:07:23.768014 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9753e530-5b55-4951-af42-71297ce5c944" containerName="nova-api-log" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.768122 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="9753e530-5b55-4951-af42-71297ce5c944" containerName="nova-api-log" Dec 27 08:07:23 crc kubenswrapper[4934]: E1227 08:07:23.768555 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9753e530-5b55-4951-af42-71297ce5c944" containerName="nova-api-api" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.768618 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="9753e530-5b55-4951-af42-71297ce5c944" containerName="nova-api-api" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.768915 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="9753e530-5b55-4951-af42-71297ce5c944" containerName="nova-api-api" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.768995 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="9753e530-5b55-4951-af42-71297ce5c944" containerName="nova-api-log" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.770448 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.774838 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.775214 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.778893 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.792449 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.834727 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-logs\") pod \"nova-api-0\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " pod="openstack/nova-api-0" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.834765 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " pod="openstack/nova-api-0" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.834851 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " pod="openstack/nova-api-0" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.834974 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-config-data\") pod \"nova-api-0\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " pod="openstack/nova-api-0" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.834997 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-public-tls-certs\") pod \"nova-api-0\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " pod="openstack/nova-api-0" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.835157 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9svgv\" (UniqueName: \"kubernetes.io/projected/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-kube-api-access-9svgv\") pod \"nova-api-0\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " pod="openstack/nova-api-0" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.937343 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " pod="openstack/nova-api-0" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.937810 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-config-data\") pod \"nova-api-0\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " pod="openstack/nova-api-0" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.938000 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-public-tls-certs\") pod \"nova-api-0\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " pod="openstack/nova-api-0" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.938953 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9svgv\" (UniqueName: \"kubernetes.io/projected/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-kube-api-access-9svgv\") pod \"nova-api-0\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " pod="openstack/nova-api-0" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.939914 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-logs\") pod \"nova-api-0\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " pod="openstack/nova-api-0" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.940223 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " pod="openstack/nova-api-0" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.941811 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-logs\") pod \"nova-api-0\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " pod="openstack/nova-api-0" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.942615 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-public-tls-certs\") pod \"nova-api-0\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " pod="openstack/nova-api-0" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.943266 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-config-data\") pod \"nova-api-0\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " pod="openstack/nova-api-0" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.944535 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " pod="openstack/nova-api-0" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.947952 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " pod="openstack/nova-api-0" Dec 27 08:07:23 crc kubenswrapper[4934]: I1227 08:07:23.957513 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9svgv\" (UniqueName: \"kubernetes.io/projected/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-kube-api-access-9svgv\") pod \"nova-api-0\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " pod="openstack/nova-api-0" Dec 27 08:07:24 crc kubenswrapper[4934]: I1227 08:07:24.085785 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:24 crc kubenswrapper[4934]: I1227 08:07:24.098768 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 27 08:07:24 crc kubenswrapper[4934]: I1227 08:07:24.106630 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:24 crc kubenswrapper[4934]: W1227 08:07:24.652888 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a2a0ae2_eafc_4d50_8a3c_9ce0dd888c8a.slice/crio-634bd64fd5fee66f7b5678624fc60f0a1631bf9238e73c58874fa32fcd83abcc WatchSource:0}: Error finding container 634bd64fd5fee66f7b5678624fc60f0a1631bf9238e73c58874fa32fcd83abcc: Status 404 returned error can't find the container with id 634bd64fd5fee66f7b5678624fc60f0a1631bf9238e73c58874fa32fcd83abcc Dec 27 08:07:24 crc kubenswrapper[4934]: I1227 08:07:24.658558 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 27 08:07:24 crc kubenswrapper[4934]: I1227 08:07:24.687431 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a","Type":"ContainerStarted","Data":"634bd64fd5fee66f7b5678624fc60f0a1631bf9238e73c58874fa32fcd83abcc"} Dec 27 08:07:24 crc kubenswrapper[4934]: I1227 08:07:24.691583 4934 generic.go:334] "Generic (PLEG): container finished" podID="8c109ea6-23b9-426e-a44b-44592c85b112" containerID="cd86a75c8fcf7377d8218e9ec6470b11b88cf7f8fa277455d763e25e1421efc6" exitCode=2 Dec 27 08:07:24 crc kubenswrapper[4934]: I1227 08:07:24.691613 4934 generic.go:334] "Generic (PLEG): container finished" podID="8c109ea6-23b9-426e-a44b-44592c85b112" containerID="f90732ae6832c4b43f77d0cb5f281af5d28022f942eb806736fda794504c2ea0" exitCode=0 Dec 27 08:07:24 crc kubenswrapper[4934]: I1227 08:07:24.691612 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c109ea6-23b9-426e-a44b-44592c85b112","Type":"ContainerDied","Data":"cd86a75c8fcf7377d8218e9ec6470b11b88cf7f8fa277455d763e25e1421efc6"} Dec 27 08:07:24 crc kubenswrapper[4934]: I1227 08:07:24.691644 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c109ea6-23b9-426e-a44b-44592c85b112","Type":"ContainerDied","Data":"f90732ae6832c4b43f77d0cb5f281af5d28022f942eb806736fda794504c2ea0"} Dec 27 08:07:24 crc kubenswrapper[4934]: I1227 08:07:24.710009 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 27 08:07:24 crc kubenswrapper[4934]: I1227 08:07:24.828195 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 27 08:07:24 crc kubenswrapper[4934]: I1227 08:07:24.828248 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 27 08:07:24 crc kubenswrapper[4934]: I1227 08:07:24.889280 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-plls4"] Dec 27 08:07:24 crc kubenswrapper[4934]: I1227 08:07:24.891035 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-plls4" Dec 27 08:07:24 crc kubenswrapper[4934]: I1227 08:07:24.896511 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 27 08:07:24 crc kubenswrapper[4934]: I1227 08:07:24.896638 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 27 08:07:24 crc kubenswrapper[4934]: I1227 08:07:24.918905 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-plls4"] Dec 27 08:07:24 crc kubenswrapper[4934]: I1227 08:07:24.967452 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-config-data\") pod \"nova-cell1-cell-mapping-plls4\" (UID: \"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e\") " pod="openstack/nova-cell1-cell-mapping-plls4" Dec 27 08:07:24 crc kubenswrapper[4934]: I1227 08:07:24.967515 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vg8h\" (UniqueName: \"kubernetes.io/projected/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-kube-api-access-9vg8h\") pod \"nova-cell1-cell-mapping-plls4\" (UID: \"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e\") " pod="openstack/nova-cell1-cell-mapping-plls4" Dec 27 08:07:24 crc kubenswrapper[4934]: I1227 08:07:24.967554 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-scripts\") pod \"nova-cell1-cell-mapping-plls4\" (UID: \"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e\") " pod="openstack/nova-cell1-cell-mapping-plls4" Dec 27 08:07:24 crc kubenswrapper[4934]: I1227 08:07:24.967669 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-plls4\" (UID: \"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e\") " pod="openstack/nova-cell1-cell-mapping-plls4" Dec 27 08:07:25 crc kubenswrapper[4934]: I1227 08:07:25.069957 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-plls4\" (UID: \"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e\") " pod="openstack/nova-cell1-cell-mapping-plls4" Dec 27 08:07:25 crc kubenswrapper[4934]: I1227 08:07:25.070313 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-config-data\") pod \"nova-cell1-cell-mapping-plls4\" (UID: \"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e\") " pod="openstack/nova-cell1-cell-mapping-plls4" Dec 27 08:07:25 crc kubenswrapper[4934]: I1227 08:07:25.070366 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vg8h\" (UniqueName: \"kubernetes.io/projected/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-kube-api-access-9vg8h\") pod \"nova-cell1-cell-mapping-plls4\" (UID: \"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e\") " pod="openstack/nova-cell1-cell-mapping-plls4" Dec 27 08:07:25 crc kubenswrapper[4934]: I1227 08:07:25.070398 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-scripts\") pod \"nova-cell1-cell-mapping-plls4\" (UID: \"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e\") " pod="openstack/nova-cell1-cell-mapping-plls4" Dec 27 08:07:25 crc kubenswrapper[4934]: I1227 08:07:25.073848 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-scripts\") pod \"nova-cell1-cell-mapping-plls4\" (UID: \"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e\") " pod="openstack/nova-cell1-cell-mapping-plls4" Dec 27 08:07:25 crc kubenswrapper[4934]: I1227 08:07:25.074150 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-plls4\" (UID: \"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e\") " pod="openstack/nova-cell1-cell-mapping-plls4" Dec 27 08:07:25 crc kubenswrapper[4934]: I1227 08:07:25.074233 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-config-data\") pod \"nova-cell1-cell-mapping-plls4\" (UID: \"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e\") " pod="openstack/nova-cell1-cell-mapping-plls4" Dec 27 08:07:25 crc kubenswrapper[4934]: I1227 08:07:25.092709 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vg8h\" (UniqueName: \"kubernetes.io/projected/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-kube-api-access-9vg8h\") pod \"nova-cell1-cell-mapping-plls4\" (UID: \"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e\") " pod="openstack/nova-cell1-cell-mapping-plls4" Dec 27 08:07:25 crc kubenswrapper[4934]: I1227 08:07:25.219020 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-plls4" Dec 27 08:07:25 crc kubenswrapper[4934]: I1227 08:07:25.480311 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9753e530-5b55-4951-af42-71297ce5c944" path="/var/lib/kubelet/pods/9753e530-5b55-4951-af42-71297ce5c944/volumes" Dec 27 08:07:25 crc kubenswrapper[4934]: I1227 08:07:25.708253 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a","Type":"ContainerStarted","Data":"5e3b9d2fb7de6226fa7a5d901db9c2807fd8607ea9fa71790dd8569eb515716f"} Dec 27 08:07:25 crc kubenswrapper[4934]: I1227 08:07:25.708306 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a","Type":"ContainerStarted","Data":"c0c31da434b1ce189f08e101194eae35187765b6d7563a33ef2f0909328ac05b"} Dec 27 08:07:25 crc kubenswrapper[4934]: W1227 08:07:25.716606 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b85e56c_313b_4bac_b1d8_bd5d7c23ba6e.slice/crio-a275a6323df2c14599d8bdba89c933c24f97f2bb5a7799e919be981693a04b6d WatchSource:0}: Error finding container a275a6323df2c14599d8bdba89c933c24f97f2bb5a7799e919be981693a04b6d: Status 404 returned error can't find the container with id a275a6323df2c14599d8bdba89c933c24f97f2bb5a7799e919be981693a04b6d Dec 27 08:07:25 crc kubenswrapper[4934]: I1227 08:07:25.726238 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-plls4"] Dec 27 08:07:25 crc kubenswrapper[4934]: I1227 08:07:25.734363 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.734342621 podStartE2EDuration="2.734342621s" podCreationTimestamp="2025-12-27 08:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:07:25.728827513 +0000 UTC m=+1506.549268107" watchObservedRunningTime="2025-12-27 08:07:25.734342621 +0000 UTC m=+1506.554783215" Dec 27 08:07:25 crc kubenswrapper[4934]: I1227 08:07:25.836264 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="97621094-4d07-4ed0-bf11-82bf65dde763" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.248:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 08:07:25 crc kubenswrapper[4934]: I1227 08:07:25.836283 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="97621094-4d07-4ed0-bf11-82bf65dde763" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.248:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 08:07:26 crc kubenswrapper[4934]: I1227 08:07:26.727154 4934 generic.go:334] "Generic (PLEG): container finished" podID="8c109ea6-23b9-426e-a44b-44592c85b112" containerID="f413a4ac455122d4bf53aee391af087cad4883b8bc0283d8b97ff272f43b2bcb" exitCode=0 Dec 27 08:07:26 crc kubenswrapper[4934]: I1227 08:07:26.727220 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c109ea6-23b9-426e-a44b-44592c85b112","Type":"ContainerDied","Data":"f413a4ac455122d4bf53aee391af087cad4883b8bc0283d8b97ff272f43b2bcb"} Dec 27 08:07:26 crc kubenswrapper[4934]: I1227 08:07:26.730745 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-plls4" event={"ID":"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e","Type":"ContainerStarted","Data":"9a51d5582a010610a3618c11caa15ff0a3118518e903e82cfdf985039b49ef73"} Dec 27 08:07:26 crc kubenswrapper[4934]: I1227 08:07:26.730808 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-plls4" event={"ID":"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e","Type":"ContainerStarted","Data":"a275a6323df2c14599d8bdba89c933c24f97f2bb5a7799e919be981693a04b6d"} Dec 27 08:07:26 crc kubenswrapper[4934]: I1227 08:07:26.766112 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-plls4" podStartSLOduration=2.766052372 podStartE2EDuration="2.766052372s" podCreationTimestamp="2025-12-27 08:07:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:07:26.756871242 +0000 UTC m=+1507.577311876" watchObservedRunningTime="2025-12-27 08:07:26.766052372 +0000 UTC m=+1507.586493006" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.021438 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.114351 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-446hx"] Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.114878 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" podUID="d1ddee33-e9fa-4651-b7d6-d39fa59b542c" containerName="dnsmasq-dns" containerID="cri-o://e36a87cfc635581d28738c58261e1b781bc8daaa5d652e75c956ee44f66f77f8" gracePeriod=10 Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.152778 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" podUID="d1ddee33-e9fa-4651-b7d6-d39fa59b542c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.238:5353: connect: connection refused" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.720638 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.749997 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-dns-svc\") pod \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.750971 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-ovsdbserver-sb\") pod \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.751100 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wtck\" (UniqueName: \"kubernetes.io/projected/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-kube-api-access-7wtck\") pod \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.751428 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-config\") pod \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.751535 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-dns-swift-storage-0\") pod \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.751597 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-ovsdbserver-nb\") pod \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\" (UID: \"d1ddee33-e9fa-4651-b7d6-d39fa59b542c\") " Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.778490 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-kube-api-access-7wtck" (OuterVolumeSpecName: "kube-api-access-7wtck") pod "d1ddee33-e9fa-4651-b7d6-d39fa59b542c" (UID: "d1ddee33-e9fa-4651-b7d6-d39fa59b542c"). InnerVolumeSpecName "kube-api-access-7wtck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.782525 4934 generic.go:334] "Generic (PLEG): container finished" podID="d1ddee33-e9fa-4651-b7d6-d39fa59b542c" containerID="e36a87cfc635581d28738c58261e1b781bc8daaa5d652e75c956ee44f66f77f8" exitCode=0 Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.783616 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.784236 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" event={"ID":"d1ddee33-e9fa-4651-b7d6-d39fa59b542c","Type":"ContainerDied","Data":"e36a87cfc635581d28738c58261e1b781bc8daaa5d652e75c956ee44f66f77f8"} Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.784269 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-446hx" event={"ID":"d1ddee33-e9fa-4651-b7d6-d39fa59b542c","Type":"ContainerDied","Data":"bf3d59f1cfb3404d87c6c0548624865cbc071697bdacbdcad9ee17fbfc0486aa"} Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.784310 4934 scope.go:117] "RemoveContainer" containerID="e36a87cfc635581d28738c58261e1b781bc8daaa5d652e75c956ee44f66f77f8" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.841838 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-config" (OuterVolumeSpecName: "config") pod "d1ddee33-e9fa-4651-b7d6-d39fa59b542c" (UID: "d1ddee33-e9fa-4651-b7d6-d39fa59b542c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.854044 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wtck\" (UniqueName: \"kubernetes.io/projected/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-kube-api-access-7wtck\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.854081 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.867809 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d1ddee33-e9fa-4651-b7d6-d39fa59b542c" (UID: "d1ddee33-e9fa-4651-b7d6-d39fa59b542c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.879701 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d1ddee33-e9fa-4651-b7d6-d39fa59b542c" (UID: "d1ddee33-e9fa-4651-b7d6-d39fa59b542c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.888819 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d1ddee33-e9fa-4651-b7d6-d39fa59b542c" (UID: "d1ddee33-e9fa-4651-b7d6-d39fa59b542c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.894523 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d1ddee33-e9fa-4651-b7d6-d39fa59b542c" (UID: "d1ddee33-e9fa-4651-b7d6-d39fa59b542c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.904438 4934 scope.go:117] "RemoveContainer" containerID="e4d6d1872e20de0be62ae2eb22ebed27b2ce0b757517e9b1c9d6688180b9bcbd" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.924609 4934 scope.go:117] "RemoveContainer" containerID="e36a87cfc635581d28738c58261e1b781bc8daaa5d652e75c956ee44f66f77f8" Dec 27 08:07:27 crc kubenswrapper[4934]: E1227 08:07:27.924989 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e36a87cfc635581d28738c58261e1b781bc8daaa5d652e75c956ee44f66f77f8\": container with ID starting with e36a87cfc635581d28738c58261e1b781bc8daaa5d652e75c956ee44f66f77f8 not found: ID does not exist" containerID="e36a87cfc635581d28738c58261e1b781bc8daaa5d652e75c956ee44f66f77f8" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.925019 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e36a87cfc635581d28738c58261e1b781bc8daaa5d652e75c956ee44f66f77f8"} err="failed to get container status \"e36a87cfc635581d28738c58261e1b781bc8daaa5d652e75c956ee44f66f77f8\": rpc error: code = NotFound desc = could not find container \"e36a87cfc635581d28738c58261e1b781bc8daaa5d652e75c956ee44f66f77f8\": container with ID starting with e36a87cfc635581d28738c58261e1b781bc8daaa5d652e75c956ee44f66f77f8 not found: ID does not exist" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.925040 4934 scope.go:117] "RemoveContainer" containerID="e4d6d1872e20de0be62ae2eb22ebed27b2ce0b757517e9b1c9d6688180b9bcbd" Dec 27 08:07:27 crc kubenswrapper[4934]: E1227 08:07:27.925252 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4d6d1872e20de0be62ae2eb22ebed27b2ce0b757517e9b1c9d6688180b9bcbd\": container with ID starting with e4d6d1872e20de0be62ae2eb22ebed27b2ce0b757517e9b1c9d6688180b9bcbd not found: ID does not exist" containerID="e4d6d1872e20de0be62ae2eb22ebed27b2ce0b757517e9b1c9d6688180b9bcbd" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.925273 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4d6d1872e20de0be62ae2eb22ebed27b2ce0b757517e9b1c9d6688180b9bcbd"} err="failed to get container status \"e4d6d1872e20de0be62ae2eb22ebed27b2ce0b757517e9b1c9d6688180b9bcbd\": rpc error: code = NotFound desc = could not find container \"e4d6d1872e20de0be62ae2eb22ebed27b2ce0b757517e9b1c9d6688180b9bcbd\": container with ID starting with e4d6d1872e20de0be62ae2eb22ebed27b2ce0b757517e9b1c9d6688180b9bcbd not found: ID does not exist" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.955965 4934 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.956004 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.956017 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:27 crc kubenswrapper[4934]: I1227 08:07:27.956027 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d1ddee33-e9fa-4651-b7d6-d39fa59b542c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:28 crc kubenswrapper[4934]: I1227 08:07:28.120177 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-446hx"] Dec 27 08:07:28 crc kubenswrapper[4934]: I1227 08:07:28.129919 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-446hx"] Dec 27 08:07:29 crc kubenswrapper[4934]: I1227 08:07:29.491551 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1ddee33-e9fa-4651-b7d6-d39fa59b542c" path="/var/lib/kubelet/pods/d1ddee33-e9fa-4651-b7d6-d39fa59b542c/volumes" Dec 27 08:07:30 crc kubenswrapper[4934]: I1227 08:07:30.840024 4934 generic.go:334] "Generic (PLEG): container finished" podID="1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e" containerID="9a51d5582a010610a3618c11caa15ff0a3118518e903e82cfdf985039b49ef73" exitCode=0 Dec 27 08:07:30 crc kubenswrapper[4934]: I1227 08:07:30.840057 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-plls4" event={"ID":"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e","Type":"ContainerDied","Data":"9a51d5582a010610a3618c11caa15ff0a3118518e903e82cfdf985039b49ef73"} Dec 27 08:07:32 crc kubenswrapper[4934]: I1227 08:07:32.310983 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-plls4" Dec 27 08:07:32 crc kubenswrapper[4934]: I1227 08:07:32.394549 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vg8h\" (UniqueName: \"kubernetes.io/projected/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-kube-api-access-9vg8h\") pod \"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e\" (UID: \"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e\") " Dec 27 08:07:32 crc kubenswrapper[4934]: I1227 08:07:32.394585 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-scripts\") pod \"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e\" (UID: \"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e\") " Dec 27 08:07:32 crc kubenswrapper[4934]: I1227 08:07:32.395011 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-combined-ca-bundle\") pod \"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e\" (UID: \"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e\") " Dec 27 08:07:32 crc kubenswrapper[4934]: I1227 08:07:32.395346 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-config-data\") pod \"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e\" (UID: \"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e\") " Dec 27 08:07:32 crc kubenswrapper[4934]: I1227 08:07:32.402568 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-scripts" (OuterVolumeSpecName: "scripts") pod "1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e" (UID: "1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:32 crc kubenswrapper[4934]: I1227 08:07:32.402835 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-kube-api-access-9vg8h" (OuterVolumeSpecName: "kube-api-access-9vg8h") pod "1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e" (UID: "1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e"). InnerVolumeSpecName "kube-api-access-9vg8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:07:32 crc kubenswrapper[4934]: I1227 08:07:32.428449 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-config-data" (OuterVolumeSpecName: "config-data") pod "1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e" (UID: "1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:32 crc kubenswrapper[4934]: I1227 08:07:32.443849 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e" (UID: "1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:32 crc kubenswrapper[4934]: I1227 08:07:32.499292 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vg8h\" (UniqueName: \"kubernetes.io/projected/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-kube-api-access-9vg8h\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:32 crc kubenswrapper[4934]: I1227 08:07:32.499716 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:32 crc kubenswrapper[4934]: I1227 08:07:32.499735 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:32 crc kubenswrapper[4934]: I1227 08:07:32.499752 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:32 crc kubenswrapper[4934]: I1227 08:07:32.872987 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-plls4" event={"ID":"1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e","Type":"ContainerDied","Data":"a275a6323df2c14599d8bdba89c933c24f97f2bb5a7799e919be981693a04b6d"} Dec 27 08:07:32 crc kubenswrapper[4934]: I1227 08:07:32.873075 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-plls4" Dec 27 08:07:32 crc kubenswrapper[4934]: I1227 08:07:32.873149 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a275a6323df2c14599d8bdba89c933c24f97f2bb5a7799e919be981693a04b6d" Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.078340 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.078645 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a" containerName="nova-api-log" containerID="cri-o://c0c31da434b1ce189f08e101194eae35187765b6d7563a33ef2f0909328ac05b" gracePeriod=30 Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.078673 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a" containerName="nova-api-api" containerID="cri-o://5e3b9d2fb7de6226fa7a5d901db9c2807fd8607ea9fa71790dd8569eb515716f" gracePeriod=30 Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.095356 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.095591 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="933b14ea-5eba-4d47-a9b1-aead5e358154" containerName="nova-scheduler-scheduler" containerID="cri-o://a8860841c6945191b9104dd6297726f8026d1078a711851ae9e09c7c9c5015c7" gracePeriod=30 Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.106887 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.107138 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="97621094-4d07-4ed0-bf11-82bf65dde763" containerName="nova-metadata-log" containerID="cri-o://5305c4245d99ea632dccdb5b3b9f627bd571d2a06383351642303ebf8a8521b5" gracePeriod=30 Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.107231 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="97621094-4d07-4ed0-bf11-82bf65dde763" containerName="nova-metadata-metadata" containerID="cri-o://b5af56b2b2efd0e89446980bdc37b9a8cd92dc2a5d1cde92aadbe4006664bde3" gracePeriod=30 Dec 27 08:07:33 crc kubenswrapper[4934]: E1227 08:07:33.370681 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97621094_4d07_4ed0_bf11_82bf65dde763.slice/crio-conmon-5305c4245d99ea632dccdb5b3b9f627bd571d2a06383351642303ebf8a8521b5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97621094_4d07_4ed0_bf11_82bf65dde763.slice/crio-5305c4245d99ea632dccdb5b3b9f627bd571d2a06383351642303ebf8a8521b5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a2a0ae2_eafc_4d50_8a3c_9ce0dd888c8a.slice/crio-c0c31da434b1ce189f08e101194eae35187765b6d7563a33ef2f0909328ac05b.scope\": RecentStats: unable to find data in memory cache]" Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.801553 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.887220 4934 generic.go:334] "Generic (PLEG): container finished" podID="0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a" containerID="5e3b9d2fb7de6226fa7a5d901db9c2807fd8607ea9fa71790dd8569eb515716f" exitCode=0 Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.887253 4934 generic.go:334] "Generic (PLEG): container finished" podID="0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a" containerID="c0c31da434b1ce189f08e101194eae35187765b6d7563a33ef2f0909328ac05b" exitCode=143 Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.887304 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a","Type":"ContainerDied","Data":"5e3b9d2fb7de6226fa7a5d901db9c2807fd8607ea9fa71790dd8569eb515716f"} Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.887382 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a","Type":"ContainerDied","Data":"c0c31da434b1ce189f08e101194eae35187765b6d7563a33ef2f0909328ac05b"} Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.887392 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a","Type":"ContainerDied","Data":"634bd64fd5fee66f7b5678624fc60f0a1631bf9238e73c58874fa32fcd83abcc"} Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.887407 4934 scope.go:117] "RemoveContainer" containerID="5e3b9d2fb7de6226fa7a5d901db9c2807fd8607ea9fa71790dd8569eb515716f" Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.887602 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.889728 4934 generic.go:334] "Generic (PLEG): container finished" podID="97621094-4d07-4ed0-bf11-82bf65dde763" containerID="5305c4245d99ea632dccdb5b3b9f627bd571d2a06383351642303ebf8a8521b5" exitCode=143 Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.889753 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"97621094-4d07-4ed0-bf11-82bf65dde763","Type":"ContainerDied","Data":"5305c4245d99ea632dccdb5b3b9f627bd571d2a06383351642303ebf8a8521b5"} Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.913697 4934 scope.go:117] "RemoveContainer" containerID="c0c31da434b1ce189f08e101194eae35187765b6d7563a33ef2f0909328ac05b" Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.935597 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9svgv\" (UniqueName: \"kubernetes.io/projected/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-kube-api-access-9svgv\") pod \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.935699 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-internal-tls-certs\") pod \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.935832 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-public-tls-certs\") pod \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.935913 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-combined-ca-bundle\") pod \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.935970 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-logs\") pod \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.936118 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-config-data\") pod \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\" (UID: \"0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a\") " Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.936428 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-logs" (OuterVolumeSpecName: "logs") pod "0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a" (UID: "0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.936744 4934 scope.go:117] "RemoveContainer" containerID="5e3b9d2fb7de6226fa7a5d901db9c2807fd8607ea9fa71790dd8569eb515716f" Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.937041 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-logs\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:33 crc kubenswrapper[4934]: E1227 08:07:33.937246 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e3b9d2fb7de6226fa7a5d901db9c2807fd8607ea9fa71790dd8569eb515716f\": container with ID starting with 5e3b9d2fb7de6226fa7a5d901db9c2807fd8607ea9fa71790dd8569eb515716f not found: ID does not exist" containerID="5e3b9d2fb7de6226fa7a5d901db9c2807fd8607ea9fa71790dd8569eb515716f" Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.937367 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e3b9d2fb7de6226fa7a5d901db9c2807fd8607ea9fa71790dd8569eb515716f"} err="failed to get container status \"5e3b9d2fb7de6226fa7a5d901db9c2807fd8607ea9fa71790dd8569eb515716f\": rpc error: code = NotFound desc = could not find container \"5e3b9d2fb7de6226fa7a5d901db9c2807fd8607ea9fa71790dd8569eb515716f\": container with ID starting with 5e3b9d2fb7de6226fa7a5d901db9c2807fd8607ea9fa71790dd8569eb515716f not found: ID does not exist" Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.937478 4934 scope.go:117] "RemoveContainer" containerID="c0c31da434b1ce189f08e101194eae35187765b6d7563a33ef2f0909328ac05b" Dec 27 08:07:33 crc kubenswrapper[4934]: E1227 08:07:33.937909 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0c31da434b1ce189f08e101194eae35187765b6d7563a33ef2f0909328ac05b\": container with ID starting with c0c31da434b1ce189f08e101194eae35187765b6d7563a33ef2f0909328ac05b not found: ID does not exist" containerID="c0c31da434b1ce189f08e101194eae35187765b6d7563a33ef2f0909328ac05b" Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.937935 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0c31da434b1ce189f08e101194eae35187765b6d7563a33ef2f0909328ac05b"} err="failed to get container status \"c0c31da434b1ce189f08e101194eae35187765b6d7563a33ef2f0909328ac05b\": rpc error: code = NotFound desc = could not find container \"c0c31da434b1ce189f08e101194eae35187765b6d7563a33ef2f0909328ac05b\": container with ID starting with c0c31da434b1ce189f08e101194eae35187765b6d7563a33ef2f0909328ac05b not found: ID does not exist" Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.937955 4934 scope.go:117] "RemoveContainer" containerID="5e3b9d2fb7de6226fa7a5d901db9c2807fd8607ea9fa71790dd8569eb515716f" Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.938299 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e3b9d2fb7de6226fa7a5d901db9c2807fd8607ea9fa71790dd8569eb515716f"} err="failed to get container status \"5e3b9d2fb7de6226fa7a5d901db9c2807fd8607ea9fa71790dd8569eb515716f\": rpc error: code = NotFound desc = could not find container \"5e3b9d2fb7de6226fa7a5d901db9c2807fd8607ea9fa71790dd8569eb515716f\": container with ID starting with 5e3b9d2fb7de6226fa7a5d901db9c2807fd8607ea9fa71790dd8569eb515716f not found: ID does not exist" Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.938323 4934 scope.go:117] "RemoveContainer" containerID="c0c31da434b1ce189f08e101194eae35187765b6d7563a33ef2f0909328ac05b" Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.938518 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0c31da434b1ce189f08e101194eae35187765b6d7563a33ef2f0909328ac05b"} err="failed to get container status \"c0c31da434b1ce189f08e101194eae35187765b6d7563a33ef2f0909328ac05b\": rpc error: code = NotFound desc = could not find container \"c0c31da434b1ce189f08e101194eae35187765b6d7563a33ef2f0909328ac05b\": container with ID starting with c0c31da434b1ce189f08e101194eae35187765b6d7563a33ef2f0909328ac05b not found: ID does not exist" Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.941351 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-kube-api-access-9svgv" (OuterVolumeSpecName: "kube-api-access-9svgv") pod "0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a" (UID: "0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a"). InnerVolumeSpecName "kube-api-access-9svgv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.966538 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-config-data" (OuterVolumeSpecName: "config-data") pod "0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a" (UID: "0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:33 crc kubenswrapper[4934]: I1227 08:07:33.971349 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a" (UID: "0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.000601 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a" (UID: "0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.010228 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a" (UID: "0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.038745 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9svgv\" (UniqueName: \"kubernetes.io/projected/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-kube-api-access-9svgv\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.038775 4934 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.038784 4934 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.038792 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.038801 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.273226 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.282955 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.299983 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 27 08:07:34 crc kubenswrapper[4934]: E1227 08:07:34.300466 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a" containerName="nova-api-log" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.300484 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a" containerName="nova-api-log" Dec 27 08:07:34 crc kubenswrapper[4934]: E1227 08:07:34.300513 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e" containerName="nova-manage" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.300520 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e" containerName="nova-manage" Dec 27 08:07:34 crc kubenswrapper[4934]: E1227 08:07:34.300536 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1ddee33-e9fa-4651-b7d6-d39fa59b542c" containerName="dnsmasq-dns" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.300541 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1ddee33-e9fa-4651-b7d6-d39fa59b542c" containerName="dnsmasq-dns" Dec 27 08:07:34 crc kubenswrapper[4934]: E1227 08:07:34.300551 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a" containerName="nova-api-api" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.300557 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a" containerName="nova-api-api" Dec 27 08:07:34 crc kubenswrapper[4934]: E1227 08:07:34.300580 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1ddee33-e9fa-4651-b7d6-d39fa59b542c" containerName="init" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.300586 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1ddee33-e9fa-4651-b7d6-d39fa59b542c" containerName="init" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.300784 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a" containerName="nova-api-log" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.300813 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a" containerName="nova-api-api" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.300828 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1ddee33-e9fa-4651-b7d6-d39fa59b542c" containerName="dnsmasq-dns" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.300840 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e" containerName="nova-manage" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.301974 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.305475 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.305506 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.311170 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.327114 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.348953 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stwsq\" (UniqueName: \"kubernetes.io/projected/22a24b14-793e-404c-9d2e-ac30a2092cab-kube-api-access-stwsq\") pod \"nova-api-0\" (UID: \"22a24b14-793e-404c-9d2e-ac30a2092cab\") " pod="openstack/nova-api-0" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.349124 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22a24b14-793e-404c-9d2e-ac30a2092cab-public-tls-certs\") pod \"nova-api-0\" (UID: \"22a24b14-793e-404c-9d2e-ac30a2092cab\") " pod="openstack/nova-api-0" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.349144 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22a24b14-793e-404c-9d2e-ac30a2092cab-config-data\") pod \"nova-api-0\" (UID: \"22a24b14-793e-404c-9d2e-ac30a2092cab\") " pod="openstack/nova-api-0" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.349172 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22a24b14-793e-404c-9d2e-ac30a2092cab-logs\") pod \"nova-api-0\" (UID: \"22a24b14-793e-404c-9d2e-ac30a2092cab\") " pod="openstack/nova-api-0" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.349222 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22a24b14-793e-404c-9d2e-ac30a2092cab-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"22a24b14-793e-404c-9d2e-ac30a2092cab\") " pod="openstack/nova-api-0" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.349257 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22a24b14-793e-404c-9d2e-ac30a2092cab-internal-tls-certs\") pod \"nova-api-0\" (UID: \"22a24b14-793e-404c-9d2e-ac30a2092cab\") " pod="openstack/nova-api-0" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.451368 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22a24b14-793e-404c-9d2e-ac30a2092cab-internal-tls-certs\") pod \"nova-api-0\" (UID: \"22a24b14-793e-404c-9d2e-ac30a2092cab\") " pod="openstack/nova-api-0" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.451496 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stwsq\" (UniqueName: \"kubernetes.io/projected/22a24b14-793e-404c-9d2e-ac30a2092cab-kube-api-access-stwsq\") pod \"nova-api-0\" (UID: \"22a24b14-793e-404c-9d2e-ac30a2092cab\") " pod="openstack/nova-api-0" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.451609 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22a24b14-793e-404c-9d2e-ac30a2092cab-public-tls-certs\") pod \"nova-api-0\" (UID: \"22a24b14-793e-404c-9d2e-ac30a2092cab\") " pod="openstack/nova-api-0" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.451628 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22a24b14-793e-404c-9d2e-ac30a2092cab-config-data\") pod \"nova-api-0\" (UID: \"22a24b14-793e-404c-9d2e-ac30a2092cab\") " pod="openstack/nova-api-0" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.451655 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22a24b14-793e-404c-9d2e-ac30a2092cab-logs\") pod \"nova-api-0\" (UID: \"22a24b14-793e-404c-9d2e-ac30a2092cab\") " pod="openstack/nova-api-0" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.451721 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22a24b14-793e-404c-9d2e-ac30a2092cab-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"22a24b14-793e-404c-9d2e-ac30a2092cab\") " pod="openstack/nova-api-0" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.452219 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/22a24b14-793e-404c-9d2e-ac30a2092cab-logs\") pod \"nova-api-0\" (UID: \"22a24b14-793e-404c-9d2e-ac30a2092cab\") " pod="openstack/nova-api-0" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.458769 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/22a24b14-793e-404c-9d2e-ac30a2092cab-config-data\") pod \"nova-api-0\" (UID: \"22a24b14-793e-404c-9d2e-ac30a2092cab\") " pod="openstack/nova-api-0" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.489817 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/22a24b14-793e-404c-9d2e-ac30a2092cab-public-tls-certs\") pod \"nova-api-0\" (UID: \"22a24b14-793e-404c-9d2e-ac30a2092cab\") " pod="openstack/nova-api-0" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.492705 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22a24b14-793e-404c-9d2e-ac30a2092cab-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"22a24b14-793e-404c-9d2e-ac30a2092cab\") " pod="openstack/nova-api-0" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.509584 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/22a24b14-793e-404c-9d2e-ac30a2092cab-internal-tls-certs\") pod \"nova-api-0\" (UID: \"22a24b14-793e-404c-9d2e-ac30a2092cab\") " pod="openstack/nova-api-0" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.517622 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stwsq\" (UniqueName: \"kubernetes.io/projected/22a24b14-793e-404c-9d2e-ac30a2092cab-kube-api-access-stwsq\") pod \"nova-api-0\" (UID: \"22a24b14-793e-404c-9d2e-ac30a2092cab\") " pod="openstack/nova-api-0" Dec 27 08:07:34 crc kubenswrapper[4934]: I1227 08:07:34.618197 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 27 08:07:35 crc kubenswrapper[4934]: W1227 08:07:35.140971 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22a24b14_793e_404c_9d2e_ac30a2092cab.slice/crio-ad2e92febd17c409fe96c02c7e41b33bbfd3ab29fdee9e8c87e8561da52e2d53 WatchSource:0}: Error finding container ad2e92febd17c409fe96c02c7e41b33bbfd3ab29fdee9e8c87e8561da52e2d53: Status 404 returned error can't find the container with id ad2e92febd17c409fe96c02c7e41b33bbfd3ab29fdee9e8c87e8561da52e2d53 Dec 27 08:07:35 crc kubenswrapper[4934]: I1227 08:07:35.141552 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 27 08:07:35 crc kubenswrapper[4934]: I1227 08:07:35.483677 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a" path="/var/lib/kubelet/pods/0a2a0ae2-eafc-4d50-8a3c-9ce0dd888c8a/volumes" Dec 27 08:07:35 crc kubenswrapper[4934]: I1227 08:07:35.920169 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22a24b14-793e-404c-9d2e-ac30a2092cab","Type":"ContainerStarted","Data":"c30aeb85398430a2397afa0843cf0664aa65afd9323352b90645ef5d088c0a64"} Dec 27 08:07:35 crc kubenswrapper[4934]: I1227 08:07:35.920208 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22a24b14-793e-404c-9d2e-ac30a2092cab","Type":"ContainerStarted","Data":"c8867a182a432fef198176e7ef0c6ca70ad808e535ff3a909db411d18923b9be"} Dec 27 08:07:35 crc kubenswrapper[4934]: I1227 08:07:35.920218 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"22a24b14-793e-404c-9d2e-ac30a2092cab","Type":"ContainerStarted","Data":"ad2e92febd17c409fe96c02c7e41b33bbfd3ab29fdee9e8c87e8561da52e2d53"} Dec 27 08:07:35 crc kubenswrapper[4934]: I1227 08:07:35.951266 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.9512435940000001 podStartE2EDuration="1.951243594s" podCreationTimestamp="2025-12-27 08:07:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:07:35.943934781 +0000 UTC m=+1516.764375375" watchObservedRunningTime="2025-12-27 08:07:35.951243594 +0000 UTC m=+1516.771684178" Dec 27 08:07:36 crc kubenswrapper[4934]: I1227 08:07:36.814644 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 27 08:07:36 crc kubenswrapper[4934]: I1227 08:07:36.920935 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97621094-4d07-4ed0-bf11-82bf65dde763-combined-ca-bundle\") pod \"97621094-4d07-4ed0-bf11-82bf65dde763\" (UID: \"97621094-4d07-4ed0-bf11-82bf65dde763\") " Dec 27 08:07:36 crc kubenswrapper[4934]: I1227 08:07:36.921217 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/97621094-4d07-4ed0-bf11-82bf65dde763-nova-metadata-tls-certs\") pod \"97621094-4d07-4ed0-bf11-82bf65dde763\" (UID: \"97621094-4d07-4ed0-bf11-82bf65dde763\") " Dec 27 08:07:36 crc kubenswrapper[4934]: I1227 08:07:36.921239 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rl6ln\" (UniqueName: \"kubernetes.io/projected/97621094-4d07-4ed0-bf11-82bf65dde763-kube-api-access-rl6ln\") pod \"97621094-4d07-4ed0-bf11-82bf65dde763\" (UID: \"97621094-4d07-4ed0-bf11-82bf65dde763\") " Dec 27 08:07:36 crc kubenswrapper[4934]: I1227 08:07:36.921333 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97621094-4d07-4ed0-bf11-82bf65dde763-config-data\") pod \"97621094-4d07-4ed0-bf11-82bf65dde763\" (UID: \"97621094-4d07-4ed0-bf11-82bf65dde763\") " Dec 27 08:07:36 crc kubenswrapper[4934]: I1227 08:07:36.921351 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97621094-4d07-4ed0-bf11-82bf65dde763-logs\") pod \"97621094-4d07-4ed0-bf11-82bf65dde763\" (UID: \"97621094-4d07-4ed0-bf11-82bf65dde763\") " Dec 27 08:07:36 crc kubenswrapper[4934]: I1227 08:07:36.922416 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97621094-4d07-4ed0-bf11-82bf65dde763-logs" (OuterVolumeSpecName: "logs") pod "97621094-4d07-4ed0-bf11-82bf65dde763" (UID: "97621094-4d07-4ed0-bf11-82bf65dde763"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:07:36 crc kubenswrapper[4934]: I1227 08:07:36.929248 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97621094-4d07-4ed0-bf11-82bf65dde763-kube-api-access-rl6ln" (OuterVolumeSpecName: "kube-api-access-rl6ln") pod "97621094-4d07-4ed0-bf11-82bf65dde763" (UID: "97621094-4d07-4ed0-bf11-82bf65dde763"). InnerVolumeSpecName "kube-api-access-rl6ln". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:07:36 crc kubenswrapper[4934]: I1227 08:07:36.936489 4934 generic.go:334] "Generic (PLEG): container finished" podID="97621094-4d07-4ed0-bf11-82bf65dde763" containerID="b5af56b2b2efd0e89446980bdc37b9a8cd92dc2a5d1cde92aadbe4006664bde3" exitCode=0 Dec 27 08:07:36 crc kubenswrapper[4934]: I1227 08:07:36.936565 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 27 08:07:36 crc kubenswrapper[4934]: I1227 08:07:36.936551 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"97621094-4d07-4ed0-bf11-82bf65dde763","Type":"ContainerDied","Data":"b5af56b2b2efd0e89446980bdc37b9a8cd92dc2a5d1cde92aadbe4006664bde3"} Dec 27 08:07:36 crc kubenswrapper[4934]: I1227 08:07:36.936621 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"97621094-4d07-4ed0-bf11-82bf65dde763","Type":"ContainerDied","Data":"c8a4abf616f5e08edc844bf7a42aacbb2f87fbe013e378d98d4f07cea658345b"} Dec 27 08:07:36 crc kubenswrapper[4934]: I1227 08:07:36.936648 4934 scope.go:117] "RemoveContainer" containerID="b5af56b2b2efd0e89446980bdc37b9a8cd92dc2a5d1cde92aadbe4006664bde3" Dec 27 08:07:36 crc kubenswrapper[4934]: I1227 08:07:36.974655 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97621094-4d07-4ed0-bf11-82bf65dde763-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "97621094-4d07-4ed0-bf11-82bf65dde763" (UID: "97621094-4d07-4ed0-bf11-82bf65dde763"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:36 crc kubenswrapper[4934]: I1227 08:07:36.979825 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97621094-4d07-4ed0-bf11-82bf65dde763-config-data" (OuterVolumeSpecName: "config-data") pod "97621094-4d07-4ed0-bf11-82bf65dde763" (UID: "97621094-4d07-4ed0-bf11-82bf65dde763"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.025399 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rl6ln\" (UniqueName: \"kubernetes.io/projected/97621094-4d07-4ed0-bf11-82bf65dde763-kube-api-access-rl6ln\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.025438 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97621094-4d07-4ed0-bf11-82bf65dde763-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.025452 4934 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97621094-4d07-4ed0-bf11-82bf65dde763-logs\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.025464 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97621094-4d07-4ed0-bf11-82bf65dde763-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.039306 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97621094-4d07-4ed0-bf11-82bf65dde763-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "97621094-4d07-4ed0-bf11-82bf65dde763" (UID: "97621094-4d07-4ed0-bf11-82bf65dde763"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.069677 4934 scope.go:117] "RemoveContainer" containerID="5305c4245d99ea632dccdb5b3b9f627bd571d2a06383351642303ebf8a8521b5" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.095201 4934 scope.go:117] "RemoveContainer" containerID="b5af56b2b2efd0e89446980bdc37b9a8cd92dc2a5d1cde92aadbe4006664bde3" Dec 27 08:07:37 crc kubenswrapper[4934]: E1227 08:07:37.095623 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5af56b2b2efd0e89446980bdc37b9a8cd92dc2a5d1cde92aadbe4006664bde3\": container with ID starting with b5af56b2b2efd0e89446980bdc37b9a8cd92dc2a5d1cde92aadbe4006664bde3 not found: ID does not exist" containerID="b5af56b2b2efd0e89446980bdc37b9a8cd92dc2a5d1cde92aadbe4006664bde3" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.095662 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5af56b2b2efd0e89446980bdc37b9a8cd92dc2a5d1cde92aadbe4006664bde3"} err="failed to get container status \"b5af56b2b2efd0e89446980bdc37b9a8cd92dc2a5d1cde92aadbe4006664bde3\": rpc error: code = NotFound desc = could not find container \"b5af56b2b2efd0e89446980bdc37b9a8cd92dc2a5d1cde92aadbe4006664bde3\": container with ID starting with b5af56b2b2efd0e89446980bdc37b9a8cd92dc2a5d1cde92aadbe4006664bde3 not found: ID does not exist" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.095687 4934 scope.go:117] "RemoveContainer" containerID="5305c4245d99ea632dccdb5b3b9f627bd571d2a06383351642303ebf8a8521b5" Dec 27 08:07:37 crc kubenswrapper[4934]: E1227 08:07:37.096126 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5305c4245d99ea632dccdb5b3b9f627bd571d2a06383351642303ebf8a8521b5\": container with ID starting with 5305c4245d99ea632dccdb5b3b9f627bd571d2a06383351642303ebf8a8521b5 not found: ID does not exist" containerID="5305c4245d99ea632dccdb5b3b9f627bd571d2a06383351642303ebf8a8521b5" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.096157 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5305c4245d99ea632dccdb5b3b9f627bd571d2a06383351642303ebf8a8521b5"} err="failed to get container status \"5305c4245d99ea632dccdb5b3b9f627bd571d2a06383351642303ebf8a8521b5\": rpc error: code = NotFound desc = could not find container \"5305c4245d99ea632dccdb5b3b9f627bd571d2a06383351642303ebf8a8521b5\": container with ID starting with 5305c4245d99ea632dccdb5b3b9f627bd571d2a06383351642303ebf8a8521b5 not found: ID does not exist" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.128136 4934 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/97621094-4d07-4ed0-bf11-82bf65dde763-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.294054 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.307598 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.330289 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 27 08:07:37 crc kubenswrapper[4934]: E1227 08:07:37.330892 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97621094-4d07-4ed0-bf11-82bf65dde763" containerName="nova-metadata-metadata" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.330908 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="97621094-4d07-4ed0-bf11-82bf65dde763" containerName="nova-metadata-metadata" Dec 27 08:07:37 crc kubenswrapper[4934]: E1227 08:07:37.330927 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97621094-4d07-4ed0-bf11-82bf65dde763" containerName="nova-metadata-log" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.330936 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="97621094-4d07-4ed0-bf11-82bf65dde763" containerName="nova-metadata-log" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.331237 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="97621094-4d07-4ed0-bf11-82bf65dde763" containerName="nova-metadata-metadata" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.331286 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="97621094-4d07-4ed0-bf11-82bf65dde763" containerName="nova-metadata-log" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.332568 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.332648 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.350812 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.351291 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 27 08:07:37 crc kubenswrapper[4934]: E1227 08:07:37.431645 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a8860841c6945191b9104dd6297726f8026d1078a711851ae9e09c7c9c5015c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 27 08:07:37 crc kubenswrapper[4934]: E1227 08:07:37.433518 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a8860841c6945191b9104dd6297726f8026d1078a711851ae9e09c7c9c5015c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 27 08:07:37 crc kubenswrapper[4934]: E1227 08:07:37.435358 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a8860841c6945191b9104dd6297726f8026d1078a711851ae9e09c7c9c5015c7" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 27 08:07:37 crc kubenswrapper[4934]: E1227 08:07:37.435448 4934 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="933b14ea-5eba-4d47-a9b1-aead5e358154" containerName="nova-scheduler-scheduler" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.450808 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8094d1d3-d139-4889-84aa-2ba422a145db-logs\") pod \"nova-metadata-0\" (UID: \"8094d1d3-d139-4889-84aa-2ba422a145db\") " pod="openstack/nova-metadata-0" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.450859 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fxwp\" (UniqueName: \"kubernetes.io/projected/8094d1d3-d139-4889-84aa-2ba422a145db-kube-api-access-2fxwp\") pod \"nova-metadata-0\" (UID: \"8094d1d3-d139-4889-84aa-2ba422a145db\") " pod="openstack/nova-metadata-0" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.450903 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8094d1d3-d139-4889-84aa-2ba422a145db-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8094d1d3-d139-4889-84aa-2ba422a145db\") " pod="openstack/nova-metadata-0" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.451019 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8094d1d3-d139-4889-84aa-2ba422a145db-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8094d1d3-d139-4889-84aa-2ba422a145db\") " pod="openstack/nova-metadata-0" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.451167 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8094d1d3-d139-4889-84aa-2ba422a145db-config-data\") pod \"nova-metadata-0\" (UID: \"8094d1d3-d139-4889-84aa-2ba422a145db\") " pod="openstack/nova-metadata-0" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.485991 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97621094-4d07-4ed0-bf11-82bf65dde763" path="/var/lib/kubelet/pods/97621094-4d07-4ed0-bf11-82bf65dde763/volumes" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.553511 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8094d1d3-d139-4889-84aa-2ba422a145db-config-data\") pod \"nova-metadata-0\" (UID: \"8094d1d3-d139-4889-84aa-2ba422a145db\") " pod="openstack/nova-metadata-0" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.553637 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8094d1d3-d139-4889-84aa-2ba422a145db-logs\") pod \"nova-metadata-0\" (UID: \"8094d1d3-d139-4889-84aa-2ba422a145db\") " pod="openstack/nova-metadata-0" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.553671 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fxwp\" (UniqueName: \"kubernetes.io/projected/8094d1d3-d139-4889-84aa-2ba422a145db-kube-api-access-2fxwp\") pod \"nova-metadata-0\" (UID: \"8094d1d3-d139-4889-84aa-2ba422a145db\") " pod="openstack/nova-metadata-0" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.553853 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8094d1d3-d139-4889-84aa-2ba422a145db-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8094d1d3-d139-4889-84aa-2ba422a145db\") " pod="openstack/nova-metadata-0" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.554142 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8094d1d3-d139-4889-84aa-2ba422a145db-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8094d1d3-d139-4889-84aa-2ba422a145db\") " pod="openstack/nova-metadata-0" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.554165 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8094d1d3-d139-4889-84aa-2ba422a145db-logs\") pod \"nova-metadata-0\" (UID: \"8094d1d3-d139-4889-84aa-2ba422a145db\") " pod="openstack/nova-metadata-0" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.558623 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8094d1d3-d139-4889-84aa-2ba422a145db-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8094d1d3-d139-4889-84aa-2ba422a145db\") " pod="openstack/nova-metadata-0" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.562846 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8094d1d3-d139-4889-84aa-2ba422a145db-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8094d1d3-d139-4889-84aa-2ba422a145db\") " pod="openstack/nova-metadata-0" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.563600 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8094d1d3-d139-4889-84aa-2ba422a145db-config-data\") pod \"nova-metadata-0\" (UID: \"8094d1d3-d139-4889-84aa-2ba422a145db\") " pod="openstack/nova-metadata-0" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.569330 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fxwp\" (UniqueName: \"kubernetes.io/projected/8094d1d3-d139-4889-84aa-2ba422a145db-kube-api-access-2fxwp\") pod \"nova-metadata-0\" (UID: \"8094d1d3-d139-4889-84aa-2ba422a145db\") " pod="openstack/nova-metadata-0" Dec 27 08:07:37 crc kubenswrapper[4934]: I1227 08:07:37.662501 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 27 08:07:38 crc kubenswrapper[4934]: I1227 08:07:38.330557 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 27 08:07:38 crc kubenswrapper[4934]: W1227 08:07:38.342345 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8094d1d3_d139_4889_84aa_2ba422a145db.slice/crio-ab83db15d42c195e0595e865503a5d3605ba13d2137a69193b626eadd3d24795 WatchSource:0}: Error finding container ab83db15d42c195e0595e865503a5d3605ba13d2137a69193b626eadd3d24795: Status 404 returned error can't find the container with id ab83db15d42c195e0595e865503a5d3605ba13d2137a69193b626eadd3d24795 Dec 27 08:07:38 crc kubenswrapper[4934]: I1227 08:07:38.850849 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 27 08:07:38 crc kubenswrapper[4934]: I1227 08:07:38.975476 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8094d1d3-d139-4889-84aa-2ba422a145db","Type":"ContainerStarted","Data":"b3b7e6a788c7afd1f135e25d23a0b9c09bae330b98b2c08921a4a939c5074d18"} Dec 27 08:07:38 crc kubenswrapper[4934]: I1227 08:07:38.975940 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8094d1d3-d139-4889-84aa-2ba422a145db","Type":"ContainerStarted","Data":"791375b6f78cd70ac156c05120c9c337c50909866992daa8f16194f387b66ac6"} Dec 27 08:07:38 crc kubenswrapper[4934]: I1227 08:07:38.975959 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8094d1d3-d139-4889-84aa-2ba422a145db","Type":"ContainerStarted","Data":"ab83db15d42c195e0595e865503a5d3605ba13d2137a69193b626eadd3d24795"} Dec 27 08:07:38 crc kubenswrapper[4934]: I1227 08:07:38.978361 4934 generic.go:334] "Generic (PLEG): container finished" podID="933b14ea-5eba-4d47-a9b1-aead5e358154" containerID="a8860841c6945191b9104dd6297726f8026d1078a711851ae9e09c7c9c5015c7" exitCode=0 Dec 27 08:07:38 crc kubenswrapper[4934]: I1227 08:07:38.978440 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"933b14ea-5eba-4d47-a9b1-aead5e358154","Type":"ContainerDied","Data":"a8860841c6945191b9104dd6297726f8026d1078a711851ae9e09c7c9c5015c7"} Dec 27 08:07:38 crc kubenswrapper[4934]: I1227 08:07:38.978482 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"933b14ea-5eba-4d47-a9b1-aead5e358154","Type":"ContainerDied","Data":"aaf175153bf72f40b031519732f510e2b908d9ff115e7866c43a6421eea177de"} Dec 27 08:07:38 crc kubenswrapper[4934]: I1227 08:07:38.978496 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 27 08:07:38 crc kubenswrapper[4934]: I1227 08:07:38.978508 4934 scope.go:117] "RemoveContainer" containerID="a8860841c6945191b9104dd6297726f8026d1078a711851ae9e09c7c9c5015c7" Dec 27 08:07:38 crc kubenswrapper[4934]: I1227 08:07:38.996916 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/933b14ea-5eba-4d47-a9b1-aead5e358154-combined-ca-bundle\") pod \"933b14ea-5eba-4d47-a9b1-aead5e358154\" (UID: \"933b14ea-5eba-4d47-a9b1-aead5e358154\") " Dec 27 08:07:38 crc kubenswrapper[4934]: I1227 08:07:38.997230 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dtx5\" (UniqueName: \"kubernetes.io/projected/933b14ea-5eba-4d47-a9b1-aead5e358154-kube-api-access-2dtx5\") pod \"933b14ea-5eba-4d47-a9b1-aead5e358154\" (UID: \"933b14ea-5eba-4d47-a9b1-aead5e358154\") " Dec 27 08:07:38 crc kubenswrapper[4934]: I1227 08:07:38.997360 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/933b14ea-5eba-4d47-a9b1-aead5e358154-config-data\") pod \"933b14ea-5eba-4d47-a9b1-aead5e358154\" (UID: \"933b14ea-5eba-4d47-a9b1-aead5e358154\") " Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.001375 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/933b14ea-5eba-4d47-a9b1-aead5e358154-kube-api-access-2dtx5" (OuterVolumeSpecName: "kube-api-access-2dtx5") pod "933b14ea-5eba-4d47-a9b1-aead5e358154" (UID: "933b14ea-5eba-4d47-a9b1-aead5e358154"). InnerVolumeSpecName "kube-api-access-2dtx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.003283 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.003235607 podStartE2EDuration="2.003235607s" podCreationTimestamp="2025-12-27 08:07:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:07:38.993312647 +0000 UTC m=+1519.813753241" watchObservedRunningTime="2025-12-27 08:07:39.003235607 +0000 UTC m=+1519.823676211" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.007894 4934 scope.go:117] "RemoveContainer" containerID="a8860841c6945191b9104dd6297726f8026d1078a711851ae9e09c7c9c5015c7" Dec 27 08:07:39 crc kubenswrapper[4934]: E1227 08:07:39.008403 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8860841c6945191b9104dd6297726f8026d1078a711851ae9e09c7c9c5015c7\": container with ID starting with a8860841c6945191b9104dd6297726f8026d1078a711851ae9e09c7c9c5015c7 not found: ID does not exist" containerID="a8860841c6945191b9104dd6297726f8026d1078a711851ae9e09c7c9c5015c7" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.008459 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8860841c6945191b9104dd6297726f8026d1078a711851ae9e09c7c9c5015c7"} err="failed to get container status \"a8860841c6945191b9104dd6297726f8026d1078a711851ae9e09c7c9c5015c7\": rpc error: code = NotFound desc = could not find container \"a8860841c6945191b9104dd6297726f8026d1078a711851ae9e09c7c9c5015c7\": container with ID starting with a8860841c6945191b9104dd6297726f8026d1078a711851ae9e09c7c9c5015c7 not found: ID does not exist" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.040231 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/933b14ea-5eba-4d47-a9b1-aead5e358154-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "933b14ea-5eba-4d47-a9b1-aead5e358154" (UID: "933b14ea-5eba-4d47-a9b1-aead5e358154"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.054038 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/933b14ea-5eba-4d47-a9b1-aead5e358154-config-data" (OuterVolumeSpecName: "config-data") pod "933b14ea-5eba-4d47-a9b1-aead5e358154" (UID: "933b14ea-5eba-4d47-a9b1-aead5e358154"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.100021 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dtx5\" (UniqueName: \"kubernetes.io/projected/933b14ea-5eba-4d47-a9b1-aead5e358154-kube-api-access-2dtx5\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.100053 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/933b14ea-5eba-4d47-a9b1-aead5e358154-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.100062 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/933b14ea-5eba-4d47-a9b1-aead5e358154-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.312815 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.339929 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.349430 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 27 08:07:39 crc kubenswrapper[4934]: E1227 08:07:39.349978 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="933b14ea-5eba-4d47-a9b1-aead5e358154" containerName="nova-scheduler-scheduler" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.350002 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="933b14ea-5eba-4d47-a9b1-aead5e358154" containerName="nova-scheduler-scheduler" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.350355 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="933b14ea-5eba-4d47-a9b1-aead5e358154" containerName="nova-scheduler-scheduler" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.351393 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.354676 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.361218 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.481104 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="933b14ea-5eba-4d47-a9b1-aead5e358154" path="/var/lib/kubelet/pods/933b14ea-5eba-4d47-a9b1-aead5e358154/volumes" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.508223 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07762cba-4cfa-43ec-b661-bf478a62d317-config-data\") pod \"nova-scheduler-0\" (UID: \"07762cba-4cfa-43ec-b661-bf478a62d317\") " pod="openstack/nova-scheduler-0" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.508277 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gmqx\" (UniqueName: \"kubernetes.io/projected/07762cba-4cfa-43ec-b661-bf478a62d317-kube-api-access-9gmqx\") pod \"nova-scheduler-0\" (UID: \"07762cba-4cfa-43ec-b661-bf478a62d317\") " pod="openstack/nova-scheduler-0" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.508339 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07762cba-4cfa-43ec-b661-bf478a62d317-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"07762cba-4cfa-43ec-b661-bf478a62d317\") " pod="openstack/nova-scheduler-0" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.610390 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07762cba-4cfa-43ec-b661-bf478a62d317-config-data\") pod \"nova-scheduler-0\" (UID: \"07762cba-4cfa-43ec-b661-bf478a62d317\") " pod="openstack/nova-scheduler-0" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.610698 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gmqx\" (UniqueName: \"kubernetes.io/projected/07762cba-4cfa-43ec-b661-bf478a62d317-kube-api-access-9gmqx\") pod \"nova-scheduler-0\" (UID: \"07762cba-4cfa-43ec-b661-bf478a62d317\") " pod="openstack/nova-scheduler-0" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.610819 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07762cba-4cfa-43ec-b661-bf478a62d317-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"07762cba-4cfa-43ec-b661-bf478a62d317\") " pod="openstack/nova-scheduler-0" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.615859 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07762cba-4cfa-43ec-b661-bf478a62d317-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"07762cba-4cfa-43ec-b661-bf478a62d317\") " pod="openstack/nova-scheduler-0" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.616566 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07762cba-4cfa-43ec-b661-bf478a62d317-config-data\") pod \"nova-scheduler-0\" (UID: \"07762cba-4cfa-43ec-b661-bf478a62d317\") " pod="openstack/nova-scheduler-0" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.628784 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gmqx\" (UniqueName: \"kubernetes.io/projected/07762cba-4cfa-43ec-b661-bf478a62d317-kube-api-access-9gmqx\") pod \"nova-scheduler-0\" (UID: \"07762cba-4cfa-43ec-b661-bf478a62d317\") " pod="openstack/nova-scheduler-0" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.729465 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.834446 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dn65z"] Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.837789 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dn65z" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.883245 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dn65z"] Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.933426 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxqfp\" (UniqueName: \"kubernetes.io/projected/c7a1968f-0dba-48a2-b289-b6ce0d21ea14-kube-api-access-pxqfp\") pod \"certified-operators-dn65z\" (UID: \"c7a1968f-0dba-48a2-b289-b6ce0d21ea14\") " pod="openshift-marketplace/certified-operators-dn65z" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.933774 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7a1968f-0dba-48a2-b289-b6ce0d21ea14-utilities\") pod \"certified-operators-dn65z\" (UID: \"c7a1968f-0dba-48a2-b289-b6ce0d21ea14\") " pod="openshift-marketplace/certified-operators-dn65z" Dec 27 08:07:39 crc kubenswrapper[4934]: I1227 08:07:39.933944 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7a1968f-0dba-48a2-b289-b6ce0d21ea14-catalog-content\") pod \"certified-operators-dn65z\" (UID: \"c7a1968f-0dba-48a2-b289-b6ce0d21ea14\") " pod="openshift-marketplace/certified-operators-dn65z" Dec 27 08:07:40 crc kubenswrapper[4934]: I1227 08:07:40.036612 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxqfp\" (UniqueName: \"kubernetes.io/projected/c7a1968f-0dba-48a2-b289-b6ce0d21ea14-kube-api-access-pxqfp\") pod \"certified-operators-dn65z\" (UID: \"c7a1968f-0dba-48a2-b289-b6ce0d21ea14\") " pod="openshift-marketplace/certified-operators-dn65z" Dec 27 08:07:40 crc kubenswrapper[4934]: I1227 08:07:40.036755 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7a1968f-0dba-48a2-b289-b6ce0d21ea14-utilities\") pod \"certified-operators-dn65z\" (UID: \"c7a1968f-0dba-48a2-b289-b6ce0d21ea14\") " pod="openshift-marketplace/certified-operators-dn65z" Dec 27 08:07:40 crc kubenswrapper[4934]: I1227 08:07:40.036814 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7a1968f-0dba-48a2-b289-b6ce0d21ea14-catalog-content\") pod \"certified-operators-dn65z\" (UID: \"c7a1968f-0dba-48a2-b289-b6ce0d21ea14\") " pod="openshift-marketplace/certified-operators-dn65z" Dec 27 08:07:40 crc kubenswrapper[4934]: I1227 08:07:40.037445 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7a1968f-0dba-48a2-b289-b6ce0d21ea14-catalog-content\") pod \"certified-operators-dn65z\" (UID: \"c7a1968f-0dba-48a2-b289-b6ce0d21ea14\") " pod="openshift-marketplace/certified-operators-dn65z" Dec 27 08:07:40 crc kubenswrapper[4934]: I1227 08:07:40.037811 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7a1968f-0dba-48a2-b289-b6ce0d21ea14-utilities\") pod \"certified-operators-dn65z\" (UID: \"c7a1968f-0dba-48a2-b289-b6ce0d21ea14\") " pod="openshift-marketplace/certified-operators-dn65z" Dec 27 08:07:40 crc kubenswrapper[4934]: I1227 08:07:40.057130 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxqfp\" (UniqueName: \"kubernetes.io/projected/c7a1968f-0dba-48a2-b289-b6ce0d21ea14-kube-api-access-pxqfp\") pod \"certified-operators-dn65z\" (UID: \"c7a1968f-0dba-48a2-b289-b6ce0d21ea14\") " pod="openshift-marketplace/certified-operators-dn65z" Dec 27 08:07:40 crc kubenswrapper[4934]: I1227 08:07:40.198857 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dn65z" Dec 27 08:07:40 crc kubenswrapper[4934]: I1227 08:07:40.259121 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 27 08:07:40 crc kubenswrapper[4934]: W1227 08:07:40.264902 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07762cba_4cfa_43ec_b661_bf478a62d317.slice/crio-3f2088f6a91f208fa4e1a3c2cf550d59545f2e8738b80d6c051658e37db6d9e0 WatchSource:0}: Error finding container 3f2088f6a91f208fa4e1a3c2cf550d59545f2e8738b80d6c051658e37db6d9e0: Status 404 returned error can't find the container with id 3f2088f6a91f208fa4e1a3c2cf550d59545f2e8738b80d6c051658e37db6d9e0 Dec 27 08:07:40 crc kubenswrapper[4934]: I1227 08:07:40.698210 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dn65z"] Dec 27 08:07:41 crc kubenswrapper[4934]: I1227 08:07:41.014732 4934 generic.go:334] "Generic (PLEG): container finished" podID="c7a1968f-0dba-48a2-b289-b6ce0d21ea14" containerID="c7d5c5a8d7211f0d9115a75682f217e3c0a1a825ef0f990b4c97b20bae7aad56" exitCode=0 Dec 27 08:07:41 crc kubenswrapper[4934]: I1227 08:07:41.014784 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn65z" event={"ID":"c7a1968f-0dba-48a2-b289-b6ce0d21ea14","Type":"ContainerDied","Data":"c7d5c5a8d7211f0d9115a75682f217e3c0a1a825ef0f990b4c97b20bae7aad56"} Dec 27 08:07:41 crc kubenswrapper[4934]: I1227 08:07:41.015107 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn65z" event={"ID":"c7a1968f-0dba-48a2-b289-b6ce0d21ea14","Type":"ContainerStarted","Data":"f015ba12569f45244dcafba503ee85499430859891502389efcfa9b0f0889718"} Dec 27 08:07:41 crc kubenswrapper[4934]: I1227 08:07:41.017323 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"07762cba-4cfa-43ec-b661-bf478a62d317","Type":"ContainerStarted","Data":"a67bb236676d933ef42ea629a2c64911311eacf8a9112d4966a66a4e2f4b4552"} Dec 27 08:07:41 crc kubenswrapper[4934]: I1227 08:07:41.017361 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"07762cba-4cfa-43ec-b661-bf478a62d317","Type":"ContainerStarted","Data":"3f2088f6a91f208fa4e1a3c2cf550d59545f2e8738b80d6c051658e37db6d9e0"} Dec 27 08:07:41 crc kubenswrapper[4934]: I1227 08:07:41.058546 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.058526954 podStartE2EDuration="2.058526954s" podCreationTimestamp="2025-12-27 08:07:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:07:41.04904184 +0000 UTC m=+1521.869482434" watchObservedRunningTime="2025-12-27 08:07:41.058526954 +0000 UTC m=+1521.878967538" Dec 27 08:07:42 crc kubenswrapper[4934]: I1227 08:07:42.663444 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 27 08:07:42 crc kubenswrapper[4934]: I1227 08:07:42.663928 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 27 08:07:43 crc kubenswrapper[4934]: I1227 08:07:43.049419 4934 generic.go:334] "Generic (PLEG): container finished" podID="1d13d0a6-ede9-45fe-9338-5a90260f0455" containerID="9ea1535deb217127b7d666092376b7a0856f76af6c706047e65a73355eafb6e4" exitCode=137 Dec 27 08:07:43 crc kubenswrapper[4934]: I1227 08:07:43.049546 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1d13d0a6-ede9-45fe-9338-5a90260f0455","Type":"ContainerDied","Data":"9ea1535deb217127b7d666092376b7a0856f76af6c706047e65a73355eafb6e4"} Dec 27 08:07:43 crc kubenswrapper[4934]: I1227 08:07:43.055039 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn65z" event={"ID":"c7a1968f-0dba-48a2-b289-b6ce0d21ea14","Type":"ContainerStarted","Data":"2975b0b382f5a27e564b2a6924bfd7c949a5d6715b8d0893e06499161f14283a"} Dec 27 08:07:43 crc kubenswrapper[4934]: I1227 08:07:43.442801 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 27 08:07:43 crc kubenswrapper[4934]: I1227 08:07:43.525117 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d13d0a6-ede9-45fe-9338-5a90260f0455-config-data\") pod \"1d13d0a6-ede9-45fe-9338-5a90260f0455\" (UID: \"1d13d0a6-ede9-45fe-9338-5a90260f0455\") " Dec 27 08:07:43 crc kubenswrapper[4934]: I1227 08:07:43.525277 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlqmv\" (UniqueName: \"kubernetes.io/projected/1d13d0a6-ede9-45fe-9338-5a90260f0455-kube-api-access-mlqmv\") pod \"1d13d0a6-ede9-45fe-9338-5a90260f0455\" (UID: \"1d13d0a6-ede9-45fe-9338-5a90260f0455\") " Dec 27 08:07:43 crc kubenswrapper[4934]: I1227 08:07:43.525369 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d13d0a6-ede9-45fe-9338-5a90260f0455-scripts\") pod \"1d13d0a6-ede9-45fe-9338-5a90260f0455\" (UID: \"1d13d0a6-ede9-45fe-9338-5a90260f0455\") " Dec 27 08:07:43 crc kubenswrapper[4934]: I1227 08:07:43.525448 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d13d0a6-ede9-45fe-9338-5a90260f0455-combined-ca-bundle\") pod \"1d13d0a6-ede9-45fe-9338-5a90260f0455\" (UID: \"1d13d0a6-ede9-45fe-9338-5a90260f0455\") " Dec 27 08:07:43 crc kubenswrapper[4934]: I1227 08:07:43.531345 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d13d0a6-ede9-45fe-9338-5a90260f0455-kube-api-access-mlqmv" (OuterVolumeSpecName: "kube-api-access-mlqmv") pod "1d13d0a6-ede9-45fe-9338-5a90260f0455" (UID: "1d13d0a6-ede9-45fe-9338-5a90260f0455"). InnerVolumeSpecName "kube-api-access-mlqmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:07:43 crc kubenswrapper[4934]: I1227 08:07:43.534726 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d13d0a6-ede9-45fe-9338-5a90260f0455-scripts" (OuterVolumeSpecName: "scripts") pod "1d13d0a6-ede9-45fe-9338-5a90260f0455" (UID: "1d13d0a6-ede9-45fe-9338-5a90260f0455"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:43 crc kubenswrapper[4934]: I1227 08:07:43.629031 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlqmv\" (UniqueName: \"kubernetes.io/projected/1d13d0a6-ede9-45fe-9338-5a90260f0455-kube-api-access-mlqmv\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:43 crc kubenswrapper[4934]: I1227 08:07:43.629058 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d13d0a6-ede9-45fe-9338-5a90260f0455-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:43 crc kubenswrapper[4934]: I1227 08:07:43.671248 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d13d0a6-ede9-45fe-9338-5a90260f0455-config-data" (OuterVolumeSpecName: "config-data") pod "1d13d0a6-ede9-45fe-9338-5a90260f0455" (UID: "1d13d0a6-ede9-45fe-9338-5a90260f0455"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:43 crc kubenswrapper[4934]: I1227 08:07:43.685820 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d13d0a6-ede9-45fe-9338-5a90260f0455-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d13d0a6-ede9-45fe-9338-5a90260f0455" (UID: "1d13d0a6-ede9-45fe-9338-5a90260f0455"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:43 crc kubenswrapper[4934]: I1227 08:07:43.730568 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d13d0a6-ede9-45fe-9338-5a90260f0455-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:43 crc kubenswrapper[4934]: I1227 08:07:43.730598 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d13d0a6-ede9-45fe-9338-5a90260f0455-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.065958 4934 generic.go:334] "Generic (PLEG): container finished" podID="c7a1968f-0dba-48a2-b289-b6ce0d21ea14" containerID="2975b0b382f5a27e564b2a6924bfd7c949a5d6715b8d0893e06499161f14283a" exitCode=0 Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.066017 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn65z" event={"ID":"c7a1968f-0dba-48a2-b289-b6ce0d21ea14","Type":"ContainerDied","Data":"2975b0b382f5a27e564b2a6924bfd7c949a5d6715b8d0893e06499161f14283a"} Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.069564 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"1d13d0a6-ede9-45fe-9338-5a90260f0455","Type":"ContainerDied","Data":"de00a807aca97606daac3aa01aa6491dca54b465662476046c10287a682e0bf9"} Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.069602 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.069635 4934 scope.go:117] "RemoveContainer" containerID="9ea1535deb217127b7d666092376b7a0856f76af6c706047e65a73355eafb6e4" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.113772 4934 scope.go:117] "RemoveContainer" containerID="e6cf52ef3f84a1b3eede7e7564cb23f48e1ed825f07acb54857b63501d13c82a" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.120213 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.144295 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.148000 4934 scope.go:117] "RemoveContainer" containerID="3acc162dd29cb480c5e3a7b72d533b66a8a7fa5367914e43729749c74ad05402" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.166247 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 27 08:07:44 crc kubenswrapper[4934]: E1227 08:07:44.166852 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d13d0a6-ede9-45fe-9338-5a90260f0455" containerName="aodh-api" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.166872 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d13d0a6-ede9-45fe-9338-5a90260f0455" containerName="aodh-api" Dec 27 08:07:44 crc kubenswrapper[4934]: E1227 08:07:44.166905 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d13d0a6-ede9-45fe-9338-5a90260f0455" containerName="aodh-listener" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.166912 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d13d0a6-ede9-45fe-9338-5a90260f0455" containerName="aodh-listener" Dec 27 08:07:44 crc kubenswrapper[4934]: E1227 08:07:44.166927 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d13d0a6-ede9-45fe-9338-5a90260f0455" containerName="aodh-notifier" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.166934 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d13d0a6-ede9-45fe-9338-5a90260f0455" containerName="aodh-notifier" Dec 27 08:07:44 crc kubenswrapper[4934]: E1227 08:07:44.166961 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d13d0a6-ede9-45fe-9338-5a90260f0455" containerName="aodh-evaluator" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.166967 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d13d0a6-ede9-45fe-9338-5a90260f0455" containerName="aodh-evaluator" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.167284 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d13d0a6-ede9-45fe-9338-5a90260f0455" containerName="aodh-listener" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.167308 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d13d0a6-ede9-45fe-9338-5a90260f0455" containerName="aodh-notifier" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.167321 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d13d0a6-ede9-45fe-9338-5a90260f0455" containerName="aodh-api" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.167343 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d13d0a6-ede9-45fe-9338-5a90260f0455" containerName="aodh-evaluator" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.170184 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.173396 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.173737 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.173737 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-wrr58" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.173745 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.174218 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.175765 4934 scope.go:117] "RemoveContainer" containerID="fd596c2c4a357feb978a9acf180d8d3ae8340b1b043c17107ff33fe257d0ce7d" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.187321 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.242750 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-public-tls-certs\") pod \"aodh-0\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " pod="openstack/aodh-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.242833 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-scripts\") pod \"aodh-0\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " pod="openstack/aodh-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.242852 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-config-data\") pod \"aodh-0\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " pod="openstack/aodh-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.242898 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5shk\" (UniqueName: \"kubernetes.io/projected/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-kube-api-access-g5shk\") pod \"aodh-0\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " pod="openstack/aodh-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.242971 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " pod="openstack/aodh-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.243092 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-internal-tls-certs\") pod \"aodh-0\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " pod="openstack/aodh-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.345031 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " pod="openstack/aodh-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.345485 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-internal-tls-certs\") pod \"aodh-0\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " pod="openstack/aodh-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.345540 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-public-tls-certs\") pod \"aodh-0\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " pod="openstack/aodh-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.345567 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-scripts\") pod \"aodh-0\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " pod="openstack/aodh-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.345586 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-config-data\") pod \"aodh-0\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " pod="openstack/aodh-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.345617 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5shk\" (UniqueName: \"kubernetes.io/projected/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-kube-api-access-g5shk\") pod \"aodh-0\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " pod="openstack/aodh-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.352711 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-internal-tls-certs\") pod \"aodh-0\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " pod="openstack/aodh-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.352738 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-scripts\") pod \"aodh-0\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " pod="openstack/aodh-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.353109 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-public-tls-certs\") pod \"aodh-0\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " pod="openstack/aodh-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.355863 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-config-data\") pod \"aodh-0\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " pod="openstack/aodh-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.358291 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-combined-ca-bundle\") pod \"aodh-0\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " pod="openstack/aodh-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.366529 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5shk\" (UniqueName: \"kubernetes.io/projected/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-kube-api-access-g5shk\") pod \"aodh-0\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " pod="openstack/aodh-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.487781 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.618822 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.619129 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 27 08:07:44 crc kubenswrapper[4934]: I1227 08:07:44.734686 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 27 08:07:45 crc kubenswrapper[4934]: I1227 08:07:45.005041 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 27 08:07:45 crc kubenswrapper[4934]: I1227 08:07:45.081266 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e4cf02a3-796a-450f-a150-5ae72e0f6f7d","Type":"ContainerStarted","Data":"8122585b52394ca06feed270d0e8e767cfe6799284964935b40bf60d6595cc13"} Dec 27 08:07:45 crc kubenswrapper[4934]: I1227 08:07:45.333665 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:07:45 crc kubenswrapper[4934]: I1227 08:07:45.333725 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:07:45 crc kubenswrapper[4934]: I1227 08:07:45.482535 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d13d0a6-ede9-45fe-9338-5a90260f0455" path="/var/lib/kubelet/pods/1d13d0a6-ede9-45fe-9338-5a90260f0455/volumes" Dec 27 08:07:45 crc kubenswrapper[4934]: I1227 08:07:45.618979 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="22a24b14-793e-404c-9d2e-ac30a2092cab" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.253:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 08:07:45 crc kubenswrapper[4934]: I1227 08:07:45.625399 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="22a24b14-793e-404c-9d2e-ac30a2092cab" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.253:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 08:07:46 crc kubenswrapper[4934]: I1227 08:07:46.125964 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn65z" event={"ID":"c7a1968f-0dba-48a2-b289-b6ce0d21ea14","Type":"ContainerStarted","Data":"cccfa1885cfc983e9e99f47daedf895b1bfbfdf358ba0137f934c0bdfcf08713"} Dec 27 08:07:46 crc kubenswrapper[4934]: I1227 08:07:46.128535 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e4cf02a3-796a-450f-a150-5ae72e0f6f7d","Type":"ContainerStarted","Data":"57c30729eec9072ee65d52f9e8e83f6cd9a51139aa6062e2adfb001dfc2bb00e"} Dec 27 08:07:46 crc kubenswrapper[4934]: I1227 08:07:46.155408 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dn65z" podStartSLOduration=3.421569344 podStartE2EDuration="7.155383066s" podCreationTimestamp="2025-12-27 08:07:39 +0000 UTC" firstStartedPulling="2025-12-27 08:07:41.017847531 +0000 UTC m=+1521.838288125" lastFinishedPulling="2025-12-27 08:07:44.751661253 +0000 UTC m=+1525.572101847" observedRunningTime="2025-12-27 08:07:46.144750506 +0000 UTC m=+1526.965191100" watchObservedRunningTime="2025-12-27 08:07:46.155383066 +0000 UTC m=+1526.975823660" Dec 27 08:07:47 crc kubenswrapper[4934]: I1227 08:07:47.143772 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e4cf02a3-796a-450f-a150-5ae72e0f6f7d","Type":"ContainerStarted","Data":"c715a325324352bc15756616883cddc8ed263dbac45506fef4fec90a2dcf9bf2"} Dec 27 08:07:47 crc kubenswrapper[4934]: I1227 08:07:47.663459 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 27 08:07:47 crc kubenswrapper[4934]: I1227 08:07:47.663852 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 27 08:07:48 crc kubenswrapper[4934]: I1227 08:07:48.232939 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e4cf02a3-796a-450f-a150-5ae72e0f6f7d","Type":"ContainerStarted","Data":"95df1e1ab27861114324155498743dc1691daf87b8293aef64c5b5f124040970"} Dec 27 08:07:48 crc kubenswrapper[4934]: I1227 08:07:48.713331 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8094d1d3-d139-4889-84aa-2ba422a145db" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.254:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 08:07:48 crc kubenswrapper[4934]: I1227 08:07:48.713369 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8094d1d3-d139-4889-84aa-2ba422a145db" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.254:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 08:07:49 crc kubenswrapper[4934]: I1227 08:07:49.155095 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="8c109ea6-23b9-426e-a44b-44592c85b112" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 27 08:07:49 crc kubenswrapper[4934]: I1227 08:07:49.246617 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e4cf02a3-796a-450f-a150-5ae72e0f6f7d","Type":"ContainerStarted","Data":"4e22b411703fbd8da59ac24c7542b39c88cde422839e82708b596306477ee2c0"} Dec 27 08:07:49 crc kubenswrapper[4934]: I1227 08:07:49.272791 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.34020417 podStartE2EDuration="5.272772151s" podCreationTimestamp="2025-12-27 08:07:44 +0000 UTC" firstStartedPulling="2025-12-27 08:07:45.018351251 +0000 UTC m=+1525.838791845" lastFinishedPulling="2025-12-27 08:07:47.950919222 +0000 UTC m=+1528.771359826" observedRunningTime="2025-12-27 08:07:49.266921459 +0000 UTC m=+1530.087362063" watchObservedRunningTime="2025-12-27 08:07:49.272772151 +0000 UTC m=+1530.093212735" Dec 27 08:07:49 crc kubenswrapper[4934]: I1227 08:07:49.731649 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 27 08:07:49 crc kubenswrapper[4934]: I1227 08:07:49.785449 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 27 08:07:50 crc kubenswrapper[4934]: I1227 08:07:50.199412 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dn65z" Dec 27 08:07:50 crc kubenswrapper[4934]: I1227 08:07:50.199477 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dn65z" Dec 27 08:07:50 crc kubenswrapper[4934]: I1227 08:07:50.301688 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 27 08:07:51 crc kubenswrapper[4934]: I1227 08:07:51.249416 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-dn65z" podUID="c7a1968f-0dba-48a2-b289-b6ce0d21ea14" containerName="registry-server" probeResult="failure" output=< Dec 27 08:07:51 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 08:07:51 crc kubenswrapper[4934]: > Dec 27 08:07:51 crc kubenswrapper[4934]: I1227 08:07:51.871065 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k84nf"] Dec 27 08:07:51 crc kubenswrapper[4934]: I1227 08:07:51.874639 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k84nf" Dec 27 08:07:51 crc kubenswrapper[4934]: I1227 08:07:51.890428 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k84nf"] Dec 27 08:07:51 crc kubenswrapper[4934]: I1227 08:07:51.966533 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b3b1180-43cd-4be2-b569-ac08a8f1d5b3-catalog-content\") pod \"redhat-operators-k84nf\" (UID: \"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3\") " pod="openshift-marketplace/redhat-operators-k84nf" Dec 27 08:07:51 crc kubenswrapper[4934]: I1227 08:07:51.966585 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2hmw\" (UniqueName: \"kubernetes.io/projected/2b3b1180-43cd-4be2-b569-ac08a8f1d5b3-kube-api-access-s2hmw\") pod \"redhat-operators-k84nf\" (UID: \"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3\") " pod="openshift-marketplace/redhat-operators-k84nf" Dec 27 08:07:51 crc kubenswrapper[4934]: I1227 08:07:51.966633 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b3b1180-43cd-4be2-b569-ac08a8f1d5b3-utilities\") pod \"redhat-operators-k84nf\" (UID: \"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3\") " pod="openshift-marketplace/redhat-operators-k84nf" Dec 27 08:07:52 crc kubenswrapper[4934]: I1227 08:07:52.068709 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b3b1180-43cd-4be2-b569-ac08a8f1d5b3-catalog-content\") pod \"redhat-operators-k84nf\" (UID: \"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3\") " pod="openshift-marketplace/redhat-operators-k84nf" Dec 27 08:07:52 crc kubenswrapper[4934]: I1227 08:07:52.068767 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2hmw\" (UniqueName: \"kubernetes.io/projected/2b3b1180-43cd-4be2-b569-ac08a8f1d5b3-kube-api-access-s2hmw\") pod \"redhat-operators-k84nf\" (UID: \"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3\") " pod="openshift-marketplace/redhat-operators-k84nf" Dec 27 08:07:52 crc kubenswrapper[4934]: I1227 08:07:52.068814 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b3b1180-43cd-4be2-b569-ac08a8f1d5b3-utilities\") pod \"redhat-operators-k84nf\" (UID: \"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3\") " pod="openshift-marketplace/redhat-operators-k84nf" Dec 27 08:07:52 crc kubenswrapper[4934]: I1227 08:07:52.069366 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b3b1180-43cd-4be2-b569-ac08a8f1d5b3-utilities\") pod \"redhat-operators-k84nf\" (UID: \"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3\") " pod="openshift-marketplace/redhat-operators-k84nf" Dec 27 08:07:52 crc kubenswrapper[4934]: I1227 08:07:52.069583 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b3b1180-43cd-4be2-b569-ac08a8f1d5b3-catalog-content\") pod \"redhat-operators-k84nf\" (UID: \"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3\") " pod="openshift-marketplace/redhat-operators-k84nf" Dec 27 08:07:52 crc kubenswrapper[4934]: I1227 08:07:52.092970 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2hmw\" (UniqueName: \"kubernetes.io/projected/2b3b1180-43cd-4be2-b569-ac08a8f1d5b3-kube-api-access-s2hmw\") pod \"redhat-operators-k84nf\" (UID: \"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3\") " pod="openshift-marketplace/redhat-operators-k84nf" Dec 27 08:07:52 crc kubenswrapper[4934]: I1227 08:07:52.264191 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k84nf" Dec 27 08:07:52 crc kubenswrapper[4934]: I1227 08:07:52.838911 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k84nf"] Dec 27 08:07:52 crc kubenswrapper[4934]: W1227 08:07:52.856811 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b3b1180_43cd_4be2_b569_ac08a8f1d5b3.slice/crio-e01556c94bdcf3adb03d2696a77bc5710256658fac30b0deb15c543595336529 WatchSource:0}: Error finding container e01556c94bdcf3adb03d2696a77bc5710256658fac30b0deb15c543595336529: Status 404 returned error can't find the container with id e01556c94bdcf3adb03d2696a77bc5710256658fac30b0deb15c543595336529 Dec 27 08:07:53 crc kubenswrapper[4934]: I1227 08:07:53.290653 4934 generic.go:334] "Generic (PLEG): container finished" podID="2b3b1180-43cd-4be2-b569-ac08a8f1d5b3" containerID="5161853f1b778e7fc603f5430949727d1ea7e4abf7581d4bd08571e1d8532f88" exitCode=0 Dec 27 08:07:53 crc kubenswrapper[4934]: I1227 08:07:53.290991 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k84nf" event={"ID":"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3","Type":"ContainerDied","Data":"5161853f1b778e7fc603f5430949727d1ea7e4abf7581d4bd08571e1d8532f88"} Dec 27 08:07:53 crc kubenswrapper[4934]: I1227 08:07:53.291022 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k84nf" event={"ID":"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3","Type":"ContainerStarted","Data":"e01556c94bdcf3adb03d2696a77bc5710256658fac30b0deb15c543595336529"} Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.240480 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.313202 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k84nf" event={"ID":"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3","Type":"ContainerStarted","Data":"549d92f05c8854c4d152d688c59b5c6f768d92ae103737d5c8a321f192cb622b"} Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.319297 4934 generic.go:334] "Generic (PLEG): container finished" podID="8c109ea6-23b9-426e-a44b-44592c85b112" containerID="d197d9cc278fca8cccff864e6cea6401f9a3bcf90c29b139b1f86282c5aeede6" exitCode=137 Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.319350 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c109ea6-23b9-426e-a44b-44592c85b112","Type":"ContainerDied","Data":"d197d9cc278fca8cccff864e6cea6401f9a3bcf90c29b139b1f86282c5aeede6"} Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.319370 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.319392 4934 scope.go:117] "RemoveContainer" containerID="d197d9cc278fca8cccff864e6cea6401f9a3bcf90c29b139b1f86282c5aeede6" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.319380 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c109ea6-23b9-426e-a44b-44592c85b112","Type":"ContainerDied","Data":"d9dc27775e1b7438b8462a95b980ea4be4d83f6ae3a8218900ccbfcfeee20ee6"} Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.331939 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-config-data\") pod \"8c109ea6-23b9-426e-a44b-44592c85b112\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.332039 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-sg-core-conf-yaml\") pod \"8c109ea6-23b9-426e-a44b-44592c85b112\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.332234 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgthd\" (UniqueName: \"kubernetes.io/projected/8c109ea6-23b9-426e-a44b-44592c85b112-kube-api-access-jgthd\") pod \"8c109ea6-23b9-426e-a44b-44592c85b112\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.332303 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-scripts\") pod \"8c109ea6-23b9-426e-a44b-44592c85b112\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.332346 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c109ea6-23b9-426e-a44b-44592c85b112-log-httpd\") pod \"8c109ea6-23b9-426e-a44b-44592c85b112\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.332507 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c109ea6-23b9-426e-a44b-44592c85b112-run-httpd\") pod \"8c109ea6-23b9-426e-a44b-44592c85b112\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.336440 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-combined-ca-bundle\") pod \"8c109ea6-23b9-426e-a44b-44592c85b112\" (UID: \"8c109ea6-23b9-426e-a44b-44592c85b112\") " Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.333025 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c109ea6-23b9-426e-a44b-44592c85b112-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8c109ea6-23b9-426e-a44b-44592c85b112" (UID: "8c109ea6-23b9-426e-a44b-44592c85b112"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.336108 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c109ea6-23b9-426e-a44b-44592c85b112-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8c109ea6-23b9-426e-a44b-44592c85b112" (UID: "8c109ea6-23b9-426e-a44b-44592c85b112"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.340445 4934 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c109ea6-23b9-426e-a44b-44592c85b112-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.340478 4934 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c109ea6-23b9-426e-a44b-44592c85b112-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.340662 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-scripts" (OuterVolumeSpecName: "scripts") pod "8c109ea6-23b9-426e-a44b-44592c85b112" (UID: "8c109ea6-23b9-426e-a44b-44592c85b112"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.354704 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c109ea6-23b9-426e-a44b-44592c85b112-kube-api-access-jgthd" (OuterVolumeSpecName: "kube-api-access-jgthd") pod "8c109ea6-23b9-426e-a44b-44592c85b112" (UID: "8c109ea6-23b9-426e-a44b-44592c85b112"). InnerVolumeSpecName "kube-api-access-jgthd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.381651 4934 scope.go:117] "RemoveContainer" containerID="cd86a75c8fcf7377d8218e9ec6470b11b88cf7f8fa277455d763e25e1421efc6" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.431708 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8c109ea6-23b9-426e-a44b-44592c85b112" (UID: "8c109ea6-23b9-426e-a44b-44592c85b112"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.445656 4934 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.445725 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgthd\" (UniqueName: \"kubernetes.io/projected/8c109ea6-23b9-426e-a44b-44592c85b112-kube-api-access-jgthd\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.445748 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.463405 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c109ea6-23b9-426e-a44b-44592c85b112" (UID: "8c109ea6-23b9-426e-a44b-44592c85b112"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.513490 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-config-data" (OuterVolumeSpecName: "config-data") pod "8c109ea6-23b9-426e-a44b-44592c85b112" (UID: "8c109ea6-23b9-426e-a44b-44592c85b112"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.525547 4934 scope.go:117] "RemoveContainer" containerID="f90732ae6832c4b43f77d0cb5f281af5d28022f942eb806736fda794504c2ea0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.548773 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.548824 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c109ea6-23b9-426e-a44b-44592c85b112-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.561931 4934 scope.go:117] "RemoveContainer" containerID="f413a4ac455122d4bf53aee391af087cad4883b8bc0283d8b97ff272f43b2bcb" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.597929 4934 scope.go:117] "RemoveContainer" containerID="d197d9cc278fca8cccff864e6cea6401f9a3bcf90c29b139b1f86282c5aeede6" Dec 27 08:07:54 crc kubenswrapper[4934]: E1227 08:07:54.598887 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d197d9cc278fca8cccff864e6cea6401f9a3bcf90c29b139b1f86282c5aeede6\": container with ID starting with d197d9cc278fca8cccff864e6cea6401f9a3bcf90c29b139b1f86282c5aeede6 not found: ID does not exist" containerID="d197d9cc278fca8cccff864e6cea6401f9a3bcf90c29b139b1f86282c5aeede6" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.598947 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d197d9cc278fca8cccff864e6cea6401f9a3bcf90c29b139b1f86282c5aeede6"} err="failed to get container status \"d197d9cc278fca8cccff864e6cea6401f9a3bcf90c29b139b1f86282c5aeede6\": rpc error: code = NotFound desc = could not find container \"d197d9cc278fca8cccff864e6cea6401f9a3bcf90c29b139b1f86282c5aeede6\": container with ID starting with d197d9cc278fca8cccff864e6cea6401f9a3bcf90c29b139b1f86282c5aeede6 not found: ID does not exist" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.598984 4934 scope.go:117] "RemoveContainer" containerID="cd86a75c8fcf7377d8218e9ec6470b11b88cf7f8fa277455d763e25e1421efc6" Dec 27 08:07:54 crc kubenswrapper[4934]: E1227 08:07:54.599499 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd86a75c8fcf7377d8218e9ec6470b11b88cf7f8fa277455d763e25e1421efc6\": container with ID starting with cd86a75c8fcf7377d8218e9ec6470b11b88cf7f8fa277455d763e25e1421efc6 not found: ID does not exist" containerID="cd86a75c8fcf7377d8218e9ec6470b11b88cf7f8fa277455d763e25e1421efc6" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.599556 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd86a75c8fcf7377d8218e9ec6470b11b88cf7f8fa277455d763e25e1421efc6"} err="failed to get container status \"cd86a75c8fcf7377d8218e9ec6470b11b88cf7f8fa277455d763e25e1421efc6\": rpc error: code = NotFound desc = could not find container \"cd86a75c8fcf7377d8218e9ec6470b11b88cf7f8fa277455d763e25e1421efc6\": container with ID starting with cd86a75c8fcf7377d8218e9ec6470b11b88cf7f8fa277455d763e25e1421efc6 not found: ID does not exist" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.599700 4934 scope.go:117] "RemoveContainer" containerID="f90732ae6832c4b43f77d0cb5f281af5d28022f942eb806736fda794504c2ea0" Dec 27 08:07:54 crc kubenswrapper[4934]: E1227 08:07:54.600020 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f90732ae6832c4b43f77d0cb5f281af5d28022f942eb806736fda794504c2ea0\": container with ID starting with f90732ae6832c4b43f77d0cb5f281af5d28022f942eb806736fda794504c2ea0 not found: ID does not exist" containerID="f90732ae6832c4b43f77d0cb5f281af5d28022f942eb806736fda794504c2ea0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.600061 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f90732ae6832c4b43f77d0cb5f281af5d28022f942eb806736fda794504c2ea0"} err="failed to get container status \"f90732ae6832c4b43f77d0cb5f281af5d28022f942eb806736fda794504c2ea0\": rpc error: code = NotFound desc = could not find container \"f90732ae6832c4b43f77d0cb5f281af5d28022f942eb806736fda794504c2ea0\": container with ID starting with f90732ae6832c4b43f77d0cb5f281af5d28022f942eb806736fda794504c2ea0 not found: ID does not exist" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.600103 4934 scope.go:117] "RemoveContainer" containerID="f413a4ac455122d4bf53aee391af087cad4883b8bc0283d8b97ff272f43b2bcb" Dec 27 08:07:54 crc kubenswrapper[4934]: E1227 08:07:54.600597 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f413a4ac455122d4bf53aee391af087cad4883b8bc0283d8b97ff272f43b2bcb\": container with ID starting with f413a4ac455122d4bf53aee391af087cad4883b8bc0283d8b97ff272f43b2bcb not found: ID does not exist" containerID="f413a4ac455122d4bf53aee391af087cad4883b8bc0283d8b97ff272f43b2bcb" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.600660 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f413a4ac455122d4bf53aee391af087cad4883b8bc0283d8b97ff272f43b2bcb"} err="failed to get container status \"f413a4ac455122d4bf53aee391af087cad4883b8bc0283d8b97ff272f43b2bcb\": rpc error: code = NotFound desc = could not find container \"f413a4ac455122d4bf53aee391af087cad4883b8bc0283d8b97ff272f43b2bcb\": container with ID starting with f413a4ac455122d4bf53aee391af087cad4883b8bc0283d8b97ff272f43b2bcb not found: ID does not exist" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.625772 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.626248 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.633312 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.635147 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.683184 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.696302 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.738881 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:07:54 crc kubenswrapper[4934]: E1227 08:07:54.739668 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c109ea6-23b9-426e-a44b-44592c85b112" containerName="ceilometer-central-agent" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.739698 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c109ea6-23b9-426e-a44b-44592c85b112" containerName="ceilometer-central-agent" Dec 27 08:07:54 crc kubenswrapper[4934]: E1227 08:07:54.739744 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c109ea6-23b9-426e-a44b-44592c85b112" containerName="proxy-httpd" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.739752 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c109ea6-23b9-426e-a44b-44592c85b112" containerName="proxy-httpd" Dec 27 08:07:54 crc kubenswrapper[4934]: E1227 08:07:54.739779 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c109ea6-23b9-426e-a44b-44592c85b112" containerName="ceilometer-notification-agent" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.739786 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c109ea6-23b9-426e-a44b-44592c85b112" containerName="ceilometer-notification-agent" Dec 27 08:07:54 crc kubenswrapper[4934]: E1227 08:07:54.739799 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c109ea6-23b9-426e-a44b-44592c85b112" containerName="sg-core" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.739805 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c109ea6-23b9-426e-a44b-44592c85b112" containerName="sg-core" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.740118 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c109ea6-23b9-426e-a44b-44592c85b112" containerName="ceilometer-central-agent" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.740158 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c109ea6-23b9-426e-a44b-44592c85b112" containerName="sg-core" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.740175 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c109ea6-23b9-426e-a44b-44592c85b112" containerName="ceilometer-notification-agent" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.740192 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c109ea6-23b9-426e-a44b-44592c85b112" containerName="proxy-httpd" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.747607 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.753694 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.753882 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.765596 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.868071 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.868195 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-scripts\") pod \"ceilometer-0\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.868226 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-config-data\") pod \"ceilometer-0\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.868253 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.868474 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5495b02-2ce5-4d67-8dbf-869d08e8891f-run-httpd\") pod \"ceilometer-0\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.868527 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5495b02-2ce5-4d67-8dbf-869d08e8891f-log-httpd\") pod \"ceilometer-0\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.868606 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgss5\" (UniqueName: \"kubernetes.io/projected/f5495b02-2ce5-4d67-8dbf-869d08e8891f-kube-api-access-tgss5\") pod \"ceilometer-0\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.970555 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgss5\" (UniqueName: \"kubernetes.io/projected/f5495b02-2ce5-4d67-8dbf-869d08e8891f-kube-api-access-tgss5\") pod \"ceilometer-0\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.970882 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.971001 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-scripts\") pod \"ceilometer-0\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.971125 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.971690 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-config-data\") pod \"ceilometer-0\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.971878 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5495b02-2ce5-4d67-8dbf-869d08e8891f-run-httpd\") pod \"ceilometer-0\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.972005 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5495b02-2ce5-4d67-8dbf-869d08e8891f-log-httpd\") pod \"ceilometer-0\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.972396 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5495b02-2ce5-4d67-8dbf-869d08e8891f-log-httpd\") pod \"ceilometer-0\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.972653 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5495b02-2ce5-4d67-8dbf-869d08e8891f-run-httpd\") pod \"ceilometer-0\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.974289 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.975500 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.975532 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-config-data\") pod \"ceilometer-0\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.976737 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-scripts\") pod \"ceilometer-0\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " pod="openstack/ceilometer-0" Dec 27 08:07:54 crc kubenswrapper[4934]: I1227 08:07:54.993310 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgss5\" (UniqueName: \"kubernetes.io/projected/f5495b02-2ce5-4d67-8dbf-869d08e8891f-kube-api-access-tgss5\") pod \"ceilometer-0\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " pod="openstack/ceilometer-0" Dec 27 08:07:55 crc kubenswrapper[4934]: I1227 08:07:55.066726 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:07:55 crc kubenswrapper[4934]: I1227 08:07:55.337665 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 27 08:07:55 crc kubenswrapper[4934]: I1227 08:07:55.363522 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 27 08:07:55 crc kubenswrapper[4934]: I1227 08:07:55.527100 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c109ea6-23b9-426e-a44b-44592c85b112" path="/var/lib/kubelet/pods/8c109ea6-23b9-426e-a44b-44592c85b112/volumes" Dec 27 08:07:55 crc kubenswrapper[4934]: I1227 08:07:55.635827 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:07:56 crc kubenswrapper[4934]: I1227 08:07:56.389145 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f5495b02-2ce5-4d67-8dbf-869d08e8891f","Type":"ContainerStarted","Data":"d04c6081111edc8d1be9d573bca9dd7023c39ac3a82676d2cb55d212d7db05d9"} Dec 27 08:07:57 crc kubenswrapper[4934]: I1227 08:07:57.403095 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f5495b02-2ce5-4d67-8dbf-869d08e8891f","Type":"ContainerStarted","Data":"3d1e128e9cd40b07a6e5b793c66fedbe9bcd9b29d9d42be19604636c8786b248"} Dec 27 08:07:57 crc kubenswrapper[4934]: I1227 08:07:57.671428 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 27 08:07:57 crc kubenswrapper[4934]: I1227 08:07:57.671956 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 27 08:07:57 crc kubenswrapper[4934]: I1227 08:07:57.682495 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 27 08:07:57 crc kubenswrapper[4934]: I1227 08:07:57.707192 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 27 08:07:58 crc kubenswrapper[4934]: I1227 08:07:58.414209 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f5495b02-2ce5-4d67-8dbf-869d08e8891f","Type":"ContainerStarted","Data":"e03df11cd32b7ae16694526d9e9ba028080674e0673db23fedec649bbf3e5289"} Dec 27 08:07:58 crc kubenswrapper[4934]: I1227 08:07:58.416858 4934 generic.go:334] "Generic (PLEG): container finished" podID="2b3b1180-43cd-4be2-b569-ac08a8f1d5b3" containerID="549d92f05c8854c4d152d688c59b5c6f768d92ae103737d5c8a321f192cb622b" exitCode=0 Dec 27 08:07:58 crc kubenswrapper[4934]: I1227 08:07:58.416885 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k84nf" event={"ID":"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3","Type":"ContainerDied","Data":"549d92f05c8854c4d152d688c59b5c6f768d92ae103737d5c8a321f192cb622b"} Dec 27 08:07:59 crc kubenswrapper[4934]: I1227 08:07:59.428593 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f5495b02-2ce5-4d67-8dbf-869d08e8891f","Type":"ContainerStarted","Data":"d5063e2de21b94886e6e47a35036b3162233d0c2588149daf93bd51d58bc367c"} Dec 27 08:07:59 crc kubenswrapper[4934]: I1227 08:07:59.432540 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k84nf" event={"ID":"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3","Type":"ContainerStarted","Data":"61394d693f3b33ddb04c67cd7b21313dd4c1c5e1427f13f47366cf5184e8aa39"} Dec 27 08:07:59 crc kubenswrapper[4934]: I1227 08:07:59.461688 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k84nf" podStartSLOduration=2.891861593 podStartE2EDuration="8.461665711s" podCreationTimestamp="2025-12-27 08:07:51 +0000 UTC" firstStartedPulling="2025-12-27 08:07:53.292535884 +0000 UTC m=+1534.112976478" lastFinishedPulling="2025-12-27 08:07:58.862340002 +0000 UTC m=+1539.682780596" observedRunningTime="2025-12-27 08:07:59.457803855 +0000 UTC m=+1540.278244449" watchObservedRunningTime="2025-12-27 08:07:59.461665711 +0000 UTC m=+1540.282106305" Dec 27 08:08:00 crc kubenswrapper[4934]: I1227 08:08:00.276258 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dn65z" Dec 27 08:08:00 crc kubenswrapper[4934]: I1227 08:08:00.408592 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dn65z" Dec 27 08:08:00 crc kubenswrapper[4934]: I1227 08:08:00.694492 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dn65z"] Dec 27 08:08:01 crc kubenswrapper[4934]: I1227 08:08:01.471469 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dn65z" podUID="c7a1968f-0dba-48a2-b289-b6ce0d21ea14" containerName="registry-server" containerID="cri-o://cccfa1885cfc983e9e99f47daedf895b1bfbfdf358ba0137f934c0bdfcf08713" gracePeriod=2 Dec 27 08:08:01 crc kubenswrapper[4934]: I1227 08:08:01.487587 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 27 08:08:01 crc kubenswrapper[4934]: I1227 08:08:01.487634 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f5495b02-2ce5-4d67-8dbf-869d08e8891f","Type":"ContainerStarted","Data":"a8454b48b28130b54f78c7bba068032994f182841bc9e8bc35ac658c7ae39d6d"} Dec 27 08:08:01 crc kubenswrapper[4934]: I1227 08:08:01.506530 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.865028029 podStartE2EDuration="7.506515384s" podCreationTimestamp="2025-12-27 08:07:54 +0000 UTC" firstStartedPulling="2025-12-27 08:07:55.639062602 +0000 UTC m=+1536.459503196" lastFinishedPulling="2025-12-27 08:08:00.280549957 +0000 UTC m=+1541.100990551" observedRunningTime="2025-12-27 08:08:01.498655549 +0000 UTC m=+1542.319096143" watchObservedRunningTime="2025-12-27 08:08:01.506515384 +0000 UTC m=+1542.326955978" Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.119754 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dn65z" Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.190298 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7a1968f-0dba-48a2-b289-b6ce0d21ea14-catalog-content\") pod \"c7a1968f-0dba-48a2-b289-b6ce0d21ea14\" (UID: \"c7a1968f-0dba-48a2-b289-b6ce0d21ea14\") " Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.190448 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxqfp\" (UniqueName: \"kubernetes.io/projected/c7a1968f-0dba-48a2-b289-b6ce0d21ea14-kube-api-access-pxqfp\") pod \"c7a1968f-0dba-48a2-b289-b6ce0d21ea14\" (UID: \"c7a1968f-0dba-48a2-b289-b6ce0d21ea14\") " Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.190504 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7a1968f-0dba-48a2-b289-b6ce0d21ea14-utilities\") pod \"c7a1968f-0dba-48a2-b289-b6ce0d21ea14\" (UID: \"c7a1968f-0dba-48a2-b289-b6ce0d21ea14\") " Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.190941 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7a1968f-0dba-48a2-b289-b6ce0d21ea14-utilities" (OuterVolumeSpecName: "utilities") pod "c7a1968f-0dba-48a2-b289-b6ce0d21ea14" (UID: "c7a1968f-0dba-48a2-b289-b6ce0d21ea14"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.202551 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7a1968f-0dba-48a2-b289-b6ce0d21ea14-kube-api-access-pxqfp" (OuterVolumeSpecName: "kube-api-access-pxqfp") pod "c7a1968f-0dba-48a2-b289-b6ce0d21ea14" (UID: "c7a1968f-0dba-48a2-b289-b6ce0d21ea14"). InnerVolumeSpecName "kube-api-access-pxqfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.252776 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7a1968f-0dba-48a2-b289-b6ce0d21ea14-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c7a1968f-0dba-48a2-b289-b6ce0d21ea14" (UID: "c7a1968f-0dba-48a2-b289-b6ce0d21ea14"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.265025 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k84nf" Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.265094 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k84nf" Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.292695 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7a1968f-0dba-48a2-b289-b6ce0d21ea14-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.292721 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxqfp\" (UniqueName: \"kubernetes.io/projected/c7a1968f-0dba-48a2-b289-b6ce0d21ea14-kube-api-access-pxqfp\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.292730 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7a1968f-0dba-48a2-b289-b6ce0d21ea14-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.495191 4934 generic.go:334] "Generic (PLEG): container finished" podID="c7a1968f-0dba-48a2-b289-b6ce0d21ea14" containerID="cccfa1885cfc983e9e99f47daedf895b1bfbfdf358ba0137f934c0bdfcf08713" exitCode=0 Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.495309 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dn65z" Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.495299 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn65z" event={"ID":"c7a1968f-0dba-48a2-b289-b6ce0d21ea14","Type":"ContainerDied","Data":"cccfa1885cfc983e9e99f47daedf895b1bfbfdf358ba0137f934c0bdfcf08713"} Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.495577 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dn65z" event={"ID":"c7a1968f-0dba-48a2-b289-b6ce0d21ea14","Type":"ContainerDied","Data":"f015ba12569f45244dcafba503ee85499430859891502389efcfa9b0f0889718"} Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.495621 4934 scope.go:117] "RemoveContainer" containerID="cccfa1885cfc983e9e99f47daedf895b1bfbfdf358ba0137f934c0bdfcf08713" Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.522165 4934 scope.go:117] "RemoveContainer" containerID="2975b0b382f5a27e564b2a6924bfd7c949a5d6715b8d0893e06499161f14283a" Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.552660 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dn65z"] Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.564928 4934 scope.go:117] "RemoveContainer" containerID="c7d5c5a8d7211f0d9115a75682f217e3c0a1a825ef0f990b4c97b20bae7aad56" Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.572453 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dn65z"] Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.614715 4934 scope.go:117] "RemoveContainer" containerID="cccfa1885cfc983e9e99f47daedf895b1bfbfdf358ba0137f934c0bdfcf08713" Dec 27 08:08:02 crc kubenswrapper[4934]: E1227 08:08:02.615347 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cccfa1885cfc983e9e99f47daedf895b1bfbfdf358ba0137f934c0bdfcf08713\": container with ID starting with cccfa1885cfc983e9e99f47daedf895b1bfbfdf358ba0137f934c0bdfcf08713 not found: ID does not exist" containerID="cccfa1885cfc983e9e99f47daedf895b1bfbfdf358ba0137f934c0bdfcf08713" Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.615407 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cccfa1885cfc983e9e99f47daedf895b1bfbfdf358ba0137f934c0bdfcf08713"} err="failed to get container status \"cccfa1885cfc983e9e99f47daedf895b1bfbfdf358ba0137f934c0bdfcf08713\": rpc error: code = NotFound desc = could not find container \"cccfa1885cfc983e9e99f47daedf895b1bfbfdf358ba0137f934c0bdfcf08713\": container with ID starting with cccfa1885cfc983e9e99f47daedf895b1bfbfdf358ba0137f934c0bdfcf08713 not found: ID does not exist" Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.615446 4934 scope.go:117] "RemoveContainer" containerID="2975b0b382f5a27e564b2a6924bfd7c949a5d6715b8d0893e06499161f14283a" Dec 27 08:08:02 crc kubenswrapper[4934]: E1227 08:08:02.615937 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2975b0b382f5a27e564b2a6924bfd7c949a5d6715b8d0893e06499161f14283a\": container with ID starting with 2975b0b382f5a27e564b2a6924bfd7c949a5d6715b8d0893e06499161f14283a not found: ID does not exist" containerID="2975b0b382f5a27e564b2a6924bfd7c949a5d6715b8d0893e06499161f14283a" Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.616167 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2975b0b382f5a27e564b2a6924bfd7c949a5d6715b8d0893e06499161f14283a"} err="failed to get container status \"2975b0b382f5a27e564b2a6924bfd7c949a5d6715b8d0893e06499161f14283a\": rpc error: code = NotFound desc = could not find container \"2975b0b382f5a27e564b2a6924bfd7c949a5d6715b8d0893e06499161f14283a\": container with ID starting with 2975b0b382f5a27e564b2a6924bfd7c949a5d6715b8d0893e06499161f14283a not found: ID does not exist" Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.616256 4934 scope.go:117] "RemoveContainer" containerID="c7d5c5a8d7211f0d9115a75682f217e3c0a1a825ef0f990b4c97b20bae7aad56" Dec 27 08:08:02 crc kubenswrapper[4934]: E1227 08:08:02.616862 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7d5c5a8d7211f0d9115a75682f217e3c0a1a825ef0f990b4c97b20bae7aad56\": container with ID starting with c7d5c5a8d7211f0d9115a75682f217e3c0a1a825ef0f990b4c97b20bae7aad56 not found: ID does not exist" containerID="c7d5c5a8d7211f0d9115a75682f217e3c0a1a825ef0f990b4c97b20bae7aad56" Dec 27 08:08:02 crc kubenswrapper[4934]: I1227 08:08:02.616907 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7d5c5a8d7211f0d9115a75682f217e3c0a1a825ef0f990b4c97b20bae7aad56"} err="failed to get container status \"c7d5c5a8d7211f0d9115a75682f217e3c0a1a825ef0f990b4c97b20bae7aad56\": rpc error: code = NotFound desc = could not find container \"c7d5c5a8d7211f0d9115a75682f217e3c0a1a825ef0f990b4c97b20bae7aad56\": container with ID starting with c7d5c5a8d7211f0d9115a75682f217e3c0a1a825ef0f990b4c97b20bae7aad56 not found: ID does not exist" Dec 27 08:08:03 crc kubenswrapper[4934]: I1227 08:08:03.329699 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k84nf" podUID="2b3b1180-43cd-4be2-b569-ac08a8f1d5b3" containerName="registry-server" probeResult="failure" output=< Dec 27 08:08:03 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 08:08:03 crc kubenswrapper[4934]: > Dec 27 08:08:03 crc kubenswrapper[4934]: I1227 08:08:03.486042 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7a1968f-0dba-48a2-b289-b6ce0d21ea14" path="/var/lib/kubelet/pods/c7a1968f-0dba-48a2-b289-b6ce0d21ea14/volumes" Dec 27 08:08:13 crc kubenswrapper[4934]: I1227 08:08:13.343686 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k84nf" podUID="2b3b1180-43cd-4be2-b569-ac08a8f1d5b3" containerName="registry-server" probeResult="failure" output=< Dec 27 08:08:13 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 08:08:13 crc kubenswrapper[4934]: > Dec 27 08:08:15 crc kubenswrapper[4934]: I1227 08:08:15.330723 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:08:15 crc kubenswrapper[4934]: I1227 08:08:15.331263 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:08:15 crc kubenswrapper[4934]: I1227 08:08:15.331350 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 08:08:15 crc kubenswrapper[4934]: I1227 08:08:15.332456 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 08:08:15 crc kubenswrapper[4934]: I1227 08:08:15.332587 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" gracePeriod=600 Dec 27 08:08:15 crc kubenswrapper[4934]: E1227 08:08:15.464793 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:08:15 crc kubenswrapper[4934]: I1227 08:08:15.705671 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" exitCode=0 Dec 27 08:08:15 crc kubenswrapper[4934]: I1227 08:08:15.705772 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab"} Dec 27 08:08:15 crc kubenswrapper[4934]: I1227 08:08:15.706130 4934 scope.go:117] "RemoveContainer" containerID="b5285e4beb4b025eb5cca462bbc9a5da90660cedc2a9f99c3f12f8c689605503" Dec 27 08:08:15 crc kubenswrapper[4934]: I1227 08:08:15.707760 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:08:15 crc kubenswrapper[4934]: E1227 08:08:15.708325 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:08:22 crc kubenswrapper[4934]: I1227 08:08:22.319400 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k84nf" Dec 27 08:08:22 crc kubenswrapper[4934]: I1227 08:08:22.365136 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k84nf" Dec 27 08:08:23 crc kubenswrapper[4934]: I1227 08:08:23.072324 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k84nf"] Dec 27 08:08:23 crc kubenswrapper[4934]: I1227 08:08:23.827867 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-k84nf" podUID="2b3b1180-43cd-4be2-b569-ac08a8f1d5b3" containerName="registry-server" containerID="cri-o://61394d693f3b33ddb04c67cd7b21313dd4c1c5e1427f13f47366cf5184e8aa39" gracePeriod=2 Dec 27 08:08:24 crc kubenswrapper[4934]: I1227 08:08:24.423288 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k84nf" Dec 27 08:08:24 crc kubenswrapper[4934]: I1227 08:08:24.580902 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b3b1180-43cd-4be2-b569-ac08a8f1d5b3-catalog-content\") pod \"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3\" (UID: \"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3\") " Dec 27 08:08:24 crc kubenswrapper[4934]: I1227 08:08:24.581374 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b3b1180-43cd-4be2-b569-ac08a8f1d5b3-utilities\") pod \"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3\" (UID: \"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3\") " Dec 27 08:08:24 crc kubenswrapper[4934]: I1227 08:08:24.581420 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2hmw\" (UniqueName: \"kubernetes.io/projected/2b3b1180-43cd-4be2-b569-ac08a8f1d5b3-kube-api-access-s2hmw\") pod \"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3\" (UID: \"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3\") " Dec 27 08:08:24 crc kubenswrapper[4934]: I1227 08:08:24.583117 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b3b1180-43cd-4be2-b569-ac08a8f1d5b3-utilities" (OuterVolumeSpecName: "utilities") pod "2b3b1180-43cd-4be2-b569-ac08a8f1d5b3" (UID: "2b3b1180-43cd-4be2-b569-ac08a8f1d5b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:08:24 crc kubenswrapper[4934]: I1227 08:08:24.591299 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b3b1180-43cd-4be2-b569-ac08a8f1d5b3-kube-api-access-s2hmw" (OuterVolumeSpecName: "kube-api-access-s2hmw") pod "2b3b1180-43cd-4be2-b569-ac08a8f1d5b3" (UID: "2b3b1180-43cd-4be2-b569-ac08a8f1d5b3"). InnerVolumeSpecName "kube-api-access-s2hmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:08:24 crc kubenswrapper[4934]: I1227 08:08:24.684610 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b3b1180-43cd-4be2-b569-ac08a8f1d5b3-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:24 crc kubenswrapper[4934]: I1227 08:08:24.684648 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2hmw\" (UniqueName: \"kubernetes.io/projected/2b3b1180-43cd-4be2-b569-ac08a8f1d5b3-kube-api-access-s2hmw\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:24 crc kubenswrapper[4934]: I1227 08:08:24.704448 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b3b1180-43cd-4be2-b569-ac08a8f1d5b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b3b1180-43cd-4be2-b569-ac08a8f1d5b3" (UID: "2b3b1180-43cd-4be2-b569-ac08a8f1d5b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:08:24 crc kubenswrapper[4934]: I1227 08:08:24.787056 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b3b1180-43cd-4be2-b569-ac08a8f1d5b3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:24 crc kubenswrapper[4934]: I1227 08:08:24.845600 4934 generic.go:334] "Generic (PLEG): container finished" podID="2b3b1180-43cd-4be2-b569-ac08a8f1d5b3" containerID="61394d693f3b33ddb04c67cd7b21313dd4c1c5e1427f13f47366cf5184e8aa39" exitCode=0 Dec 27 08:08:24 crc kubenswrapper[4934]: I1227 08:08:24.845673 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k84nf" Dec 27 08:08:24 crc kubenswrapper[4934]: I1227 08:08:24.845674 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k84nf" event={"ID":"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3","Type":"ContainerDied","Data":"61394d693f3b33ddb04c67cd7b21313dd4c1c5e1427f13f47366cf5184e8aa39"} Dec 27 08:08:24 crc kubenswrapper[4934]: I1227 08:08:24.845745 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k84nf" event={"ID":"2b3b1180-43cd-4be2-b569-ac08a8f1d5b3","Type":"ContainerDied","Data":"e01556c94bdcf3adb03d2696a77bc5710256658fac30b0deb15c543595336529"} Dec 27 08:08:24 crc kubenswrapper[4934]: I1227 08:08:24.845783 4934 scope.go:117] "RemoveContainer" containerID="61394d693f3b33ddb04c67cd7b21313dd4c1c5e1427f13f47366cf5184e8aa39" Dec 27 08:08:24 crc kubenswrapper[4934]: I1227 08:08:24.883666 4934 scope.go:117] "RemoveContainer" containerID="549d92f05c8854c4d152d688c59b5c6f768d92ae103737d5c8a321f192cb622b" Dec 27 08:08:24 crc kubenswrapper[4934]: I1227 08:08:24.904547 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k84nf"] Dec 27 08:08:24 crc kubenswrapper[4934]: I1227 08:08:24.917680 4934 scope.go:117] "RemoveContainer" containerID="5161853f1b778e7fc603f5430949727d1ea7e4abf7581d4bd08571e1d8532f88" Dec 27 08:08:24 crc kubenswrapper[4934]: I1227 08:08:24.924694 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-k84nf"] Dec 27 08:08:25 crc kubenswrapper[4934]: I1227 08:08:25.000046 4934 scope.go:117] "RemoveContainer" containerID="61394d693f3b33ddb04c67cd7b21313dd4c1c5e1427f13f47366cf5184e8aa39" Dec 27 08:08:25 crc kubenswrapper[4934]: E1227 08:08:25.000557 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61394d693f3b33ddb04c67cd7b21313dd4c1c5e1427f13f47366cf5184e8aa39\": container with ID starting with 61394d693f3b33ddb04c67cd7b21313dd4c1c5e1427f13f47366cf5184e8aa39 not found: ID does not exist" containerID="61394d693f3b33ddb04c67cd7b21313dd4c1c5e1427f13f47366cf5184e8aa39" Dec 27 08:08:25 crc kubenswrapper[4934]: I1227 08:08:25.000630 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61394d693f3b33ddb04c67cd7b21313dd4c1c5e1427f13f47366cf5184e8aa39"} err="failed to get container status \"61394d693f3b33ddb04c67cd7b21313dd4c1c5e1427f13f47366cf5184e8aa39\": rpc error: code = NotFound desc = could not find container \"61394d693f3b33ddb04c67cd7b21313dd4c1c5e1427f13f47366cf5184e8aa39\": container with ID starting with 61394d693f3b33ddb04c67cd7b21313dd4c1c5e1427f13f47366cf5184e8aa39 not found: ID does not exist" Dec 27 08:08:25 crc kubenswrapper[4934]: I1227 08:08:25.000670 4934 scope.go:117] "RemoveContainer" containerID="549d92f05c8854c4d152d688c59b5c6f768d92ae103737d5c8a321f192cb622b" Dec 27 08:08:25 crc kubenswrapper[4934]: E1227 08:08:25.001188 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"549d92f05c8854c4d152d688c59b5c6f768d92ae103737d5c8a321f192cb622b\": container with ID starting with 549d92f05c8854c4d152d688c59b5c6f768d92ae103737d5c8a321f192cb622b not found: ID does not exist" containerID="549d92f05c8854c4d152d688c59b5c6f768d92ae103737d5c8a321f192cb622b" Dec 27 08:08:25 crc kubenswrapper[4934]: I1227 08:08:25.001225 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"549d92f05c8854c4d152d688c59b5c6f768d92ae103737d5c8a321f192cb622b"} err="failed to get container status \"549d92f05c8854c4d152d688c59b5c6f768d92ae103737d5c8a321f192cb622b\": rpc error: code = NotFound desc = could not find container \"549d92f05c8854c4d152d688c59b5c6f768d92ae103737d5c8a321f192cb622b\": container with ID starting with 549d92f05c8854c4d152d688c59b5c6f768d92ae103737d5c8a321f192cb622b not found: ID does not exist" Dec 27 08:08:25 crc kubenswrapper[4934]: I1227 08:08:25.001249 4934 scope.go:117] "RemoveContainer" containerID="5161853f1b778e7fc603f5430949727d1ea7e4abf7581d4bd08571e1d8532f88" Dec 27 08:08:25 crc kubenswrapper[4934]: E1227 08:08:25.001666 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5161853f1b778e7fc603f5430949727d1ea7e4abf7581d4bd08571e1d8532f88\": container with ID starting with 5161853f1b778e7fc603f5430949727d1ea7e4abf7581d4bd08571e1d8532f88 not found: ID does not exist" containerID="5161853f1b778e7fc603f5430949727d1ea7e4abf7581d4bd08571e1d8532f88" Dec 27 08:08:25 crc kubenswrapper[4934]: I1227 08:08:25.001723 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5161853f1b778e7fc603f5430949727d1ea7e4abf7581d4bd08571e1d8532f88"} err="failed to get container status \"5161853f1b778e7fc603f5430949727d1ea7e4abf7581d4bd08571e1d8532f88\": rpc error: code = NotFound desc = could not find container \"5161853f1b778e7fc603f5430949727d1ea7e4abf7581d4bd08571e1d8532f88\": container with ID starting with 5161853f1b778e7fc603f5430949727d1ea7e4abf7581d4bd08571e1d8532f88 not found: ID does not exist" Dec 27 08:08:25 crc kubenswrapper[4934]: I1227 08:08:25.075544 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 27 08:08:25 crc kubenswrapper[4934]: I1227 08:08:25.503121 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b3b1180-43cd-4be2-b569-ac08a8f1d5b3" path="/var/lib/kubelet/pods/2b3b1180-43cd-4be2-b569-ac08a8f1d5b3/volumes" Dec 27 08:08:28 crc kubenswrapper[4934]: I1227 08:08:28.469235 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:08:28 crc kubenswrapper[4934]: E1227 08:08:28.470284 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:08:29 crc kubenswrapper[4934]: I1227 08:08:29.177322 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 27 08:08:29 crc kubenswrapper[4934]: I1227 08:08:29.177578 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="26cea6d5-e2df-46ba-a731-722d91bd3c91" containerName="kube-state-metrics" containerID="cri-o://ccfc511f4ea276acde9189dd054526355385fbd2bcc5b07b341eaf9ddbb666ba" gracePeriod=30 Dec 27 08:08:29 crc kubenswrapper[4934]: I1227 08:08:29.319031 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 27 08:08:29 crc kubenswrapper[4934]: I1227 08:08:29.319468 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mysqld-exporter-0" podUID="040f4268-549c-4c0e-885f-cc9a7dcf4e03" containerName="mysqld-exporter" containerID="cri-o://0c3314af0879f460f0bbef2e66c9195a567ab4265adce822d865320da4505fa7" gracePeriod=30 Dec 27 08:08:29 crc kubenswrapper[4934]: I1227 08:08:29.909277 4934 generic.go:334] "Generic (PLEG): container finished" podID="040f4268-549c-4c0e-885f-cc9a7dcf4e03" containerID="0c3314af0879f460f0bbef2e66c9195a567ab4265adce822d865320da4505fa7" exitCode=2 Dec 27 08:08:29 crc kubenswrapper[4934]: I1227 08:08:29.909580 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"040f4268-549c-4c0e-885f-cc9a7dcf4e03","Type":"ContainerDied","Data":"0c3314af0879f460f0bbef2e66c9195a567ab4265adce822d865320da4505fa7"} Dec 27 08:08:29 crc kubenswrapper[4934]: I1227 08:08:29.919339 4934 generic.go:334] "Generic (PLEG): container finished" podID="26cea6d5-e2df-46ba-a731-722d91bd3c91" containerID="ccfc511f4ea276acde9189dd054526355385fbd2bcc5b07b341eaf9ddbb666ba" exitCode=2 Dec 27 08:08:29 crc kubenswrapper[4934]: I1227 08:08:29.919384 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"26cea6d5-e2df-46ba-a731-722d91bd3c91","Type":"ContainerDied","Data":"ccfc511f4ea276acde9189dd054526355385fbd2bcc5b07b341eaf9ddbb666ba"} Dec 27 08:08:29 crc kubenswrapper[4934]: I1227 08:08:29.919411 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"26cea6d5-e2df-46ba-a731-722d91bd3c91","Type":"ContainerDied","Data":"c5e933b18911653b5ac892d1e3730195c01f29c6b25d58fff0297c6dca4d6db3"} Dec 27 08:08:29 crc kubenswrapper[4934]: I1227 08:08:29.919421 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5e933b18911653b5ac892d1e3730195c01f29c6b25d58fff0297c6dca4d6db3" Dec 27 08:08:30 crc kubenswrapper[4934]: I1227 08:08:30.037231 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 27 08:08:30 crc kubenswrapper[4934]: I1227 08:08:30.044556 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 27 08:08:30 crc kubenswrapper[4934]: I1227 08:08:30.234621 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/040f4268-549c-4c0e-885f-cc9a7dcf4e03-config-data\") pod \"040f4268-549c-4c0e-885f-cc9a7dcf4e03\" (UID: \"040f4268-549c-4c0e-885f-cc9a7dcf4e03\") " Dec 27 08:08:30 crc kubenswrapper[4934]: I1227 08:08:30.234816 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/040f4268-549c-4c0e-885f-cc9a7dcf4e03-combined-ca-bundle\") pod \"040f4268-549c-4c0e-885f-cc9a7dcf4e03\" (UID: \"040f4268-549c-4c0e-885f-cc9a7dcf4e03\") " Dec 27 08:08:30 crc kubenswrapper[4934]: I1227 08:08:30.234873 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lvg9\" (UniqueName: \"kubernetes.io/projected/040f4268-549c-4c0e-885f-cc9a7dcf4e03-kube-api-access-7lvg9\") pod \"040f4268-549c-4c0e-885f-cc9a7dcf4e03\" (UID: \"040f4268-549c-4c0e-885f-cc9a7dcf4e03\") " Dec 27 08:08:30 crc kubenswrapper[4934]: I1227 08:08:30.235602 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nk66k\" (UniqueName: \"kubernetes.io/projected/26cea6d5-e2df-46ba-a731-722d91bd3c91-kube-api-access-nk66k\") pod \"26cea6d5-e2df-46ba-a731-722d91bd3c91\" (UID: \"26cea6d5-e2df-46ba-a731-722d91bd3c91\") " Dec 27 08:08:30 crc kubenswrapper[4934]: I1227 08:08:30.241352 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/040f4268-549c-4c0e-885f-cc9a7dcf4e03-kube-api-access-7lvg9" (OuterVolumeSpecName: "kube-api-access-7lvg9") pod "040f4268-549c-4c0e-885f-cc9a7dcf4e03" (UID: "040f4268-549c-4c0e-885f-cc9a7dcf4e03"). InnerVolumeSpecName "kube-api-access-7lvg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:08:30 crc kubenswrapper[4934]: I1227 08:08:30.242842 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26cea6d5-e2df-46ba-a731-722d91bd3c91-kube-api-access-nk66k" (OuterVolumeSpecName: "kube-api-access-nk66k") pod "26cea6d5-e2df-46ba-a731-722d91bd3c91" (UID: "26cea6d5-e2df-46ba-a731-722d91bd3c91"). InnerVolumeSpecName "kube-api-access-nk66k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:08:30 crc kubenswrapper[4934]: I1227 08:08:30.279145 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/040f4268-549c-4c0e-885f-cc9a7dcf4e03-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "040f4268-549c-4c0e-885f-cc9a7dcf4e03" (UID: "040f4268-549c-4c0e-885f-cc9a7dcf4e03"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:08:30 crc kubenswrapper[4934]: I1227 08:08:30.305199 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/040f4268-549c-4c0e-885f-cc9a7dcf4e03-config-data" (OuterVolumeSpecName: "config-data") pod "040f4268-549c-4c0e-885f-cc9a7dcf4e03" (UID: "040f4268-549c-4c0e-885f-cc9a7dcf4e03"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:08:30 crc kubenswrapper[4934]: I1227 08:08:30.338270 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/040f4268-549c-4c0e-885f-cc9a7dcf4e03-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:30 crc kubenswrapper[4934]: I1227 08:08:30.338305 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lvg9\" (UniqueName: \"kubernetes.io/projected/040f4268-549c-4c0e-885f-cc9a7dcf4e03-kube-api-access-7lvg9\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:30 crc kubenswrapper[4934]: I1227 08:08:30.338316 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nk66k\" (UniqueName: \"kubernetes.io/projected/26cea6d5-e2df-46ba-a731-722d91bd3c91-kube-api-access-nk66k\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:30 crc kubenswrapper[4934]: I1227 08:08:30.338325 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/040f4268-549c-4c0e-885f-cc9a7dcf4e03-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:30 crc kubenswrapper[4934]: I1227 08:08:30.931467 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 27 08:08:30 crc kubenswrapper[4934]: I1227 08:08:30.931501 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 27 08:08:30 crc kubenswrapper[4934]: I1227 08:08:30.931493 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"040f4268-549c-4c0e-885f-cc9a7dcf4e03","Type":"ContainerDied","Data":"5d736ec8cff677b28af8405f06f025faf4d1c49e10784b46224ac48aa00d8a5b"} Dec 27 08:08:30 crc kubenswrapper[4934]: I1227 08:08:30.933215 4934 scope.go:117] "RemoveContainer" containerID="0c3314af0879f460f0bbef2e66c9195a567ab4265adce822d865320da4505fa7" Dec 27 08:08:30 crc kubenswrapper[4934]: I1227 08:08:30.984823 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.002287 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.018230 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.028208 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.037059 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Dec 27 08:08:31 crc kubenswrapper[4934]: E1227 08:08:31.037564 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7a1968f-0dba-48a2-b289-b6ce0d21ea14" containerName="registry-server" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.037591 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7a1968f-0dba-48a2-b289-b6ce0d21ea14" containerName="registry-server" Dec 27 08:08:31 crc kubenswrapper[4934]: E1227 08:08:31.037605 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7a1968f-0dba-48a2-b289-b6ce0d21ea14" containerName="extract-content" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.037611 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7a1968f-0dba-48a2-b289-b6ce0d21ea14" containerName="extract-content" Dec 27 08:08:31 crc kubenswrapper[4934]: E1227 08:08:31.037627 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b3b1180-43cd-4be2-b569-ac08a8f1d5b3" containerName="registry-server" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.037633 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b3b1180-43cd-4be2-b569-ac08a8f1d5b3" containerName="registry-server" Dec 27 08:08:31 crc kubenswrapper[4934]: E1227 08:08:31.037644 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="040f4268-549c-4c0e-885f-cc9a7dcf4e03" containerName="mysqld-exporter" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.037651 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="040f4268-549c-4c0e-885f-cc9a7dcf4e03" containerName="mysqld-exporter" Dec 27 08:08:31 crc kubenswrapper[4934]: E1227 08:08:31.037668 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26cea6d5-e2df-46ba-a731-722d91bd3c91" containerName="kube-state-metrics" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.037674 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="26cea6d5-e2df-46ba-a731-722d91bd3c91" containerName="kube-state-metrics" Dec 27 08:08:31 crc kubenswrapper[4934]: E1227 08:08:31.037688 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b3b1180-43cd-4be2-b569-ac08a8f1d5b3" containerName="extract-utilities" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.037695 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b3b1180-43cd-4be2-b569-ac08a8f1d5b3" containerName="extract-utilities" Dec 27 08:08:31 crc kubenswrapper[4934]: E1227 08:08:31.037705 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b3b1180-43cd-4be2-b569-ac08a8f1d5b3" containerName="extract-content" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.037711 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b3b1180-43cd-4be2-b569-ac08a8f1d5b3" containerName="extract-content" Dec 27 08:08:31 crc kubenswrapper[4934]: E1227 08:08:31.037744 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7a1968f-0dba-48a2-b289-b6ce0d21ea14" containerName="extract-utilities" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.037750 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7a1968f-0dba-48a2-b289-b6ce0d21ea14" containerName="extract-utilities" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.037941 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="26cea6d5-e2df-46ba-a731-722d91bd3c91" containerName="kube-state-metrics" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.037956 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b3b1180-43cd-4be2-b569-ac08a8f1d5b3" containerName="registry-server" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.037976 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7a1968f-0dba-48a2-b289-b6ce0d21ea14" containerName="registry-server" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.037990 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="040f4268-549c-4c0e-885f-cc9a7dcf4e03" containerName="mysqld-exporter" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.038819 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.047432 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.048921 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.051467 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-mysqld-exporter-svc" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.051543 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.051691 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.053800 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.080627 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.102948 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.154660 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf70522b-7d97-4388-b097-daf31bc4b535-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"bf70522b-7d97-4388-b097-daf31bc4b535\") " pod="openstack/mysqld-exporter-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.154801 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6jgm\" (UniqueName: \"kubernetes.io/projected/bf70522b-7d97-4388-b097-daf31bc4b535-kube-api-access-p6jgm\") pod \"mysqld-exporter-0\" (UID: \"bf70522b-7d97-4388-b097-daf31bc4b535\") " pod="openstack/mysqld-exporter-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.154928 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/4e39152f-3592-4907-9b06-d13074aeb3c4-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"4e39152f-3592-4907-9b06-d13074aeb3c4\") " pod="openstack/kube-state-metrics-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.154996 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e39152f-3592-4907-9b06-d13074aeb3c4-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"4e39152f-3592-4907-9b06-d13074aeb3c4\") " pod="openstack/kube-state-metrics-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.155022 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf70522b-7d97-4388-b097-daf31bc4b535-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"bf70522b-7d97-4388-b097-daf31bc4b535\") " pod="openstack/mysqld-exporter-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.155160 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgdbs\" (UniqueName: \"kubernetes.io/projected/4e39152f-3592-4907-9b06-d13074aeb3c4-kube-api-access-lgdbs\") pod \"kube-state-metrics-0\" (UID: \"4e39152f-3592-4907-9b06-d13074aeb3c4\") " pod="openstack/kube-state-metrics-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.155234 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf70522b-7d97-4388-b097-daf31bc4b535-config-data\") pod \"mysqld-exporter-0\" (UID: \"bf70522b-7d97-4388-b097-daf31bc4b535\") " pod="openstack/mysqld-exporter-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.155436 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e39152f-3592-4907-9b06-d13074aeb3c4-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"4e39152f-3592-4907-9b06-d13074aeb3c4\") " pod="openstack/kube-state-metrics-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.258717 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6jgm\" (UniqueName: \"kubernetes.io/projected/bf70522b-7d97-4388-b097-daf31bc4b535-kube-api-access-p6jgm\") pod \"mysqld-exporter-0\" (UID: \"bf70522b-7d97-4388-b097-daf31bc4b535\") " pod="openstack/mysqld-exporter-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.258794 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/4e39152f-3592-4907-9b06-d13074aeb3c4-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"4e39152f-3592-4907-9b06-d13074aeb3c4\") " pod="openstack/kube-state-metrics-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.258834 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e39152f-3592-4907-9b06-d13074aeb3c4-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"4e39152f-3592-4907-9b06-d13074aeb3c4\") " pod="openstack/kube-state-metrics-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.258857 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf70522b-7d97-4388-b097-daf31bc4b535-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"bf70522b-7d97-4388-b097-daf31bc4b535\") " pod="openstack/mysqld-exporter-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.258925 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgdbs\" (UniqueName: \"kubernetes.io/projected/4e39152f-3592-4907-9b06-d13074aeb3c4-kube-api-access-lgdbs\") pod \"kube-state-metrics-0\" (UID: \"4e39152f-3592-4907-9b06-d13074aeb3c4\") " pod="openstack/kube-state-metrics-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.258963 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf70522b-7d97-4388-b097-daf31bc4b535-config-data\") pod \"mysqld-exporter-0\" (UID: \"bf70522b-7d97-4388-b097-daf31bc4b535\") " pod="openstack/mysqld-exporter-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.259050 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e39152f-3592-4907-9b06-d13074aeb3c4-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"4e39152f-3592-4907-9b06-d13074aeb3c4\") " pod="openstack/kube-state-metrics-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.259139 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf70522b-7d97-4388-b097-daf31bc4b535-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"bf70522b-7d97-4388-b097-daf31bc4b535\") " pod="openstack/mysqld-exporter-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.263398 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf70522b-7d97-4388-b097-daf31bc4b535-config-data\") pod \"mysqld-exporter-0\" (UID: \"bf70522b-7d97-4388-b097-daf31bc4b535\") " pod="openstack/mysqld-exporter-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.264906 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e39152f-3592-4907-9b06-d13074aeb3c4-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"4e39152f-3592-4907-9b06-d13074aeb3c4\") " pod="openstack/kube-state-metrics-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.265619 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e39152f-3592-4907-9b06-d13074aeb3c4-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"4e39152f-3592-4907-9b06-d13074aeb3c4\") " pod="openstack/kube-state-metrics-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.268460 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf70522b-7d97-4388-b097-daf31bc4b535-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"bf70522b-7d97-4388-b097-daf31bc4b535\") " pod="openstack/mysqld-exporter-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.269700 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/4e39152f-3592-4907-9b06-d13074aeb3c4-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"4e39152f-3592-4907-9b06-d13074aeb3c4\") " pod="openstack/kube-state-metrics-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.269721 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf70522b-7d97-4388-b097-daf31bc4b535-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"bf70522b-7d97-4388-b097-daf31bc4b535\") " pod="openstack/mysqld-exporter-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.274462 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6jgm\" (UniqueName: \"kubernetes.io/projected/bf70522b-7d97-4388-b097-daf31bc4b535-kube-api-access-p6jgm\") pod \"mysqld-exporter-0\" (UID: \"bf70522b-7d97-4388-b097-daf31bc4b535\") " pod="openstack/mysqld-exporter-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.283374 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgdbs\" (UniqueName: \"kubernetes.io/projected/4e39152f-3592-4907-9b06-d13074aeb3c4-kube-api-access-lgdbs\") pod \"kube-state-metrics-0\" (UID: \"4e39152f-3592-4907-9b06-d13074aeb3c4\") " pod="openstack/kube-state-metrics-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.381128 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.390096 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.420778 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.421040 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f5495b02-2ce5-4d67-8dbf-869d08e8891f" containerName="ceilometer-central-agent" containerID="cri-o://3d1e128e9cd40b07a6e5b793c66fedbe9bcd9b29d9d42be19604636c8786b248" gracePeriod=30 Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.421161 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f5495b02-2ce5-4d67-8dbf-869d08e8891f" containerName="proxy-httpd" containerID="cri-o://a8454b48b28130b54f78c7bba068032994f182841bc9e8bc35ac658c7ae39d6d" gracePeriod=30 Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.421195 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f5495b02-2ce5-4d67-8dbf-869d08e8891f" containerName="sg-core" containerID="cri-o://d5063e2de21b94886e6e47a35036b3162233d0c2588149daf93bd51d58bc367c" gracePeriod=30 Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.421223 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f5495b02-2ce5-4d67-8dbf-869d08e8891f" containerName="ceilometer-notification-agent" containerID="cri-o://e03df11cd32b7ae16694526d9e9ba028080674e0673db23fedec649bbf3e5289" gracePeriod=30 Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.481729 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="040f4268-549c-4c0e-885f-cc9a7dcf4e03" path="/var/lib/kubelet/pods/040f4268-549c-4c0e-885f-cc9a7dcf4e03/volumes" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.482925 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26cea6d5-e2df-46ba-a731-722d91bd3c91" path="/var/lib/kubelet/pods/26cea6d5-e2df-46ba-a731-722d91bd3c91/volumes" Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.945530 4934 generic.go:334] "Generic (PLEG): container finished" podID="f5495b02-2ce5-4d67-8dbf-869d08e8891f" containerID="a8454b48b28130b54f78c7bba068032994f182841bc9e8bc35ac658c7ae39d6d" exitCode=0 Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.945854 4934 generic.go:334] "Generic (PLEG): container finished" podID="f5495b02-2ce5-4d67-8dbf-869d08e8891f" containerID="d5063e2de21b94886e6e47a35036b3162233d0c2588149daf93bd51d58bc367c" exitCode=2 Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.945865 4934 generic.go:334] "Generic (PLEG): container finished" podID="f5495b02-2ce5-4d67-8dbf-869d08e8891f" containerID="3d1e128e9cd40b07a6e5b793c66fedbe9bcd9b29d9d42be19604636c8786b248" exitCode=0 Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.945580 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f5495b02-2ce5-4d67-8dbf-869d08e8891f","Type":"ContainerDied","Data":"a8454b48b28130b54f78c7bba068032994f182841bc9e8bc35ac658c7ae39d6d"} Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.945923 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f5495b02-2ce5-4d67-8dbf-869d08e8891f","Type":"ContainerDied","Data":"d5063e2de21b94886e6e47a35036b3162233d0c2588149daf93bd51d58bc367c"} Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.945937 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f5495b02-2ce5-4d67-8dbf-869d08e8891f","Type":"ContainerDied","Data":"3d1e128e9cd40b07a6e5b793c66fedbe9bcd9b29d9d42be19604636c8786b248"} Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.956037 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-g8b74"] Dec 27 08:08:31 crc kubenswrapper[4934]: I1227 08:08:31.958557 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g8b74" Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.002144 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g8b74"] Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.036733 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.047984 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.068273 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.110140 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/784ae8a1-3984-4b1d-830d-203e05f150df-utilities\") pod \"redhat-marketplace-g8b74\" (UID: \"784ae8a1-3984-4b1d-830d-203e05f150df\") " pod="openshift-marketplace/redhat-marketplace-g8b74" Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.110623 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxm89\" (UniqueName: \"kubernetes.io/projected/784ae8a1-3984-4b1d-830d-203e05f150df-kube-api-access-rxm89\") pod \"redhat-marketplace-g8b74\" (UID: \"784ae8a1-3984-4b1d-830d-203e05f150df\") " pod="openshift-marketplace/redhat-marketplace-g8b74" Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.110779 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/784ae8a1-3984-4b1d-830d-203e05f150df-catalog-content\") pod \"redhat-marketplace-g8b74\" (UID: \"784ae8a1-3984-4b1d-830d-203e05f150df\") " pod="openshift-marketplace/redhat-marketplace-g8b74" Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.213057 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/784ae8a1-3984-4b1d-830d-203e05f150df-utilities\") pod \"redhat-marketplace-g8b74\" (UID: \"784ae8a1-3984-4b1d-830d-203e05f150df\") " pod="openshift-marketplace/redhat-marketplace-g8b74" Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.213203 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxm89\" (UniqueName: \"kubernetes.io/projected/784ae8a1-3984-4b1d-830d-203e05f150df-kube-api-access-rxm89\") pod \"redhat-marketplace-g8b74\" (UID: \"784ae8a1-3984-4b1d-830d-203e05f150df\") " pod="openshift-marketplace/redhat-marketplace-g8b74" Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.213244 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/784ae8a1-3984-4b1d-830d-203e05f150df-catalog-content\") pod \"redhat-marketplace-g8b74\" (UID: \"784ae8a1-3984-4b1d-830d-203e05f150df\") " pod="openshift-marketplace/redhat-marketplace-g8b74" Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.213761 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/784ae8a1-3984-4b1d-830d-203e05f150df-catalog-content\") pod \"redhat-marketplace-g8b74\" (UID: \"784ae8a1-3984-4b1d-830d-203e05f150df\") " pod="openshift-marketplace/redhat-marketplace-g8b74" Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.213963 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/784ae8a1-3984-4b1d-830d-203e05f150df-utilities\") pod \"redhat-marketplace-g8b74\" (UID: \"784ae8a1-3984-4b1d-830d-203e05f150df\") " pod="openshift-marketplace/redhat-marketplace-g8b74" Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.232507 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxm89\" (UniqueName: \"kubernetes.io/projected/784ae8a1-3984-4b1d-830d-203e05f150df-kube-api-access-rxm89\") pod \"redhat-marketplace-g8b74\" (UID: \"784ae8a1-3984-4b1d-830d-203e05f150df\") " pod="openshift-marketplace/redhat-marketplace-g8b74" Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.288482 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g8b74" Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.910689 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g8b74"] Dec 27 08:08:32 crc kubenswrapper[4934]: W1227 08:08:32.913438 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod784ae8a1_3984_4b1d_830d_203e05f150df.slice/crio-942a7040e8f92f86bd08615728c5008b8dbee5e9193d2253b6faebe3499fca28 WatchSource:0}: Error finding container 942a7040e8f92f86bd08615728c5008b8dbee5e9193d2253b6faebe3499fca28: Status 404 returned error can't find the container with id 942a7040e8f92f86bd08615728c5008b8dbee5e9193d2253b6faebe3499fca28 Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.968106 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g8b74" event={"ID":"784ae8a1-3984-4b1d-830d-203e05f150df","Type":"ContainerStarted","Data":"942a7040e8f92f86bd08615728c5008b8dbee5e9193d2253b6faebe3499fca28"} Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.972026 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"bf70522b-7d97-4388-b097-daf31bc4b535","Type":"ContainerStarted","Data":"659604fb90aa8503bfeeadd5e3f6e5abe0baa86084fffd62b7b74beaf682fb0c"} Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.972073 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"bf70522b-7d97-4388-b097-daf31bc4b535","Type":"ContainerStarted","Data":"814c617b8421713d38e16dae7c8472c36ffb02ff2be07dbdfe98b9ffc1cee609"} Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.982298 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4e39152f-3592-4907-9b06-d13074aeb3c4","Type":"ContainerStarted","Data":"2437eadde08e055b94e28a357e0995921db59807bc010af89911716c3aabd6e0"} Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.982360 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4e39152f-3592-4907-9b06-d13074aeb3c4","Type":"ContainerStarted","Data":"d06ebc1c7262c74ed189d00b1a58c44b9ab317038ee7cb20f4fd4765db83348b"} Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.983434 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 27 08:08:32 crc kubenswrapper[4934]: I1227 08:08:32.990969 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=2.355645678 podStartE2EDuration="2.99095241s" podCreationTimestamp="2025-12-27 08:08:30 +0000 UTC" firstStartedPulling="2025-12-27 08:08:32.061801778 +0000 UTC m=+1572.882242372" lastFinishedPulling="2025-12-27 08:08:32.69710852 +0000 UTC m=+1573.517549104" observedRunningTime="2025-12-27 08:08:32.98610306 +0000 UTC m=+1573.806543654" watchObservedRunningTime="2025-12-27 08:08:32.99095241 +0000 UTC m=+1573.811393004" Dec 27 08:08:33 crc kubenswrapper[4934]: I1227 08:08:33.006823 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.58509166 podStartE2EDuration="3.006806013s" podCreationTimestamp="2025-12-27 08:08:30 +0000 UTC" firstStartedPulling="2025-12-27 08:08:32.047594855 +0000 UTC m=+1572.868035449" lastFinishedPulling="2025-12-27 08:08:32.469309208 +0000 UTC m=+1573.289749802" observedRunningTime="2025-12-27 08:08:33.003762428 +0000 UTC m=+1573.824203022" watchObservedRunningTime="2025-12-27 08:08:33.006806013 +0000 UTC m=+1573.827246607" Dec 27 08:08:33 crc kubenswrapper[4934]: I1227 08:08:33.996808 4934 generic.go:334] "Generic (PLEG): container finished" podID="784ae8a1-3984-4b1d-830d-203e05f150df" containerID="2c92046d337fa00ce7799cb0163cc967278772f32ddfc48be0b7fa748b271665" exitCode=0 Dec 27 08:08:33 crc kubenswrapper[4934]: I1227 08:08:33.996943 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g8b74" event={"ID":"784ae8a1-3984-4b1d-830d-203e05f150df","Type":"ContainerDied","Data":"2c92046d337fa00ce7799cb0163cc967278772f32ddfc48be0b7fa748b271665"} Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.022293 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g8b74" event={"ID":"784ae8a1-3984-4b1d-830d-203e05f150df","Type":"ContainerStarted","Data":"b21f5414f614273e6d18af71037c49ddc262adbf8a2f8ac8da5ef144ee2d597d"} Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.028150 4934 generic.go:334] "Generic (PLEG): container finished" podID="f5495b02-2ce5-4d67-8dbf-869d08e8891f" containerID="e03df11cd32b7ae16694526d9e9ba028080674e0673db23fedec649bbf3e5289" exitCode=0 Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.029890 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f5495b02-2ce5-4d67-8dbf-869d08e8891f","Type":"ContainerDied","Data":"e03df11cd32b7ae16694526d9e9ba028080674e0673db23fedec649bbf3e5289"} Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.191749 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.289181 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-scripts\") pod \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.289235 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5495b02-2ce5-4d67-8dbf-869d08e8891f-run-httpd\") pod \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.289288 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-combined-ca-bundle\") pod \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.289312 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-sg-core-conf-yaml\") pod \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.289356 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-config-data\") pod \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.289503 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tgss5\" (UniqueName: \"kubernetes.io/projected/f5495b02-2ce5-4d67-8dbf-869d08e8891f-kube-api-access-tgss5\") pod \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.289594 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5495b02-2ce5-4d67-8dbf-869d08e8891f-log-httpd\") pod \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\" (UID: \"f5495b02-2ce5-4d67-8dbf-869d08e8891f\") " Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.290467 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5495b02-2ce5-4d67-8dbf-869d08e8891f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f5495b02-2ce5-4d67-8dbf-869d08e8891f" (UID: "f5495b02-2ce5-4d67-8dbf-869d08e8891f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.290490 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5495b02-2ce5-4d67-8dbf-869d08e8891f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f5495b02-2ce5-4d67-8dbf-869d08e8891f" (UID: "f5495b02-2ce5-4d67-8dbf-869d08e8891f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.296211 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5495b02-2ce5-4d67-8dbf-869d08e8891f-kube-api-access-tgss5" (OuterVolumeSpecName: "kube-api-access-tgss5") pod "f5495b02-2ce5-4d67-8dbf-869d08e8891f" (UID: "f5495b02-2ce5-4d67-8dbf-869d08e8891f"). InnerVolumeSpecName "kube-api-access-tgss5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.297502 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-scripts" (OuterVolumeSpecName: "scripts") pod "f5495b02-2ce5-4d67-8dbf-869d08e8891f" (UID: "f5495b02-2ce5-4d67-8dbf-869d08e8891f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.346093 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f5495b02-2ce5-4d67-8dbf-869d08e8891f" (UID: "f5495b02-2ce5-4d67-8dbf-869d08e8891f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.386798 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f5495b02-2ce5-4d67-8dbf-869d08e8891f" (UID: "f5495b02-2ce5-4d67-8dbf-869d08e8891f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.392423 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.392449 4934 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.392459 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tgss5\" (UniqueName: \"kubernetes.io/projected/f5495b02-2ce5-4d67-8dbf-869d08e8891f-kube-api-access-tgss5\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.392469 4934 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5495b02-2ce5-4d67-8dbf-869d08e8891f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.392477 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.392487 4934 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f5495b02-2ce5-4d67-8dbf-869d08e8891f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.432431 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-config-data" (OuterVolumeSpecName: "config-data") pod "f5495b02-2ce5-4d67-8dbf-869d08e8891f" (UID: "f5495b02-2ce5-4d67-8dbf-869d08e8891f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:08:35 crc kubenswrapper[4934]: I1227 08:08:35.494948 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5495b02-2ce5-4d67-8dbf-869d08e8891f-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.040230 4934 generic.go:334] "Generic (PLEG): container finished" podID="784ae8a1-3984-4b1d-830d-203e05f150df" containerID="b21f5414f614273e6d18af71037c49ddc262adbf8a2f8ac8da5ef144ee2d597d" exitCode=0 Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.040285 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g8b74" event={"ID":"784ae8a1-3984-4b1d-830d-203e05f150df","Type":"ContainerDied","Data":"b21f5414f614273e6d18af71037c49ddc262adbf8a2f8ac8da5ef144ee2d597d"} Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.043456 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f5495b02-2ce5-4d67-8dbf-869d08e8891f","Type":"ContainerDied","Data":"d04c6081111edc8d1be9d573bca9dd7023c39ac3a82676d2cb55d212d7db05d9"} Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.043500 4934 scope.go:117] "RemoveContainer" containerID="a8454b48b28130b54f78c7bba068032994f182841bc9e8bc35ac658c7ae39d6d" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.043588 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.103450 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.116655 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.119520 4934 scope.go:117] "RemoveContainer" containerID="d5063e2de21b94886e6e47a35036b3162233d0c2588149daf93bd51d58bc367c" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.131361 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:08:36 crc kubenswrapper[4934]: E1227 08:08:36.131980 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5495b02-2ce5-4d67-8dbf-869d08e8891f" containerName="ceilometer-notification-agent" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.132004 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5495b02-2ce5-4d67-8dbf-869d08e8891f" containerName="ceilometer-notification-agent" Dec 27 08:08:36 crc kubenswrapper[4934]: E1227 08:08:36.132034 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5495b02-2ce5-4d67-8dbf-869d08e8891f" containerName="sg-core" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.132043 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5495b02-2ce5-4d67-8dbf-869d08e8891f" containerName="sg-core" Dec 27 08:08:36 crc kubenswrapper[4934]: E1227 08:08:36.132066 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5495b02-2ce5-4d67-8dbf-869d08e8891f" containerName="ceilometer-central-agent" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.132075 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5495b02-2ce5-4d67-8dbf-869d08e8891f" containerName="ceilometer-central-agent" Dec 27 08:08:36 crc kubenswrapper[4934]: E1227 08:08:36.132307 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5495b02-2ce5-4d67-8dbf-869d08e8891f" containerName="proxy-httpd" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.132322 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5495b02-2ce5-4d67-8dbf-869d08e8891f" containerName="proxy-httpd" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.132658 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5495b02-2ce5-4d67-8dbf-869d08e8891f" containerName="sg-core" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.132702 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5495b02-2ce5-4d67-8dbf-869d08e8891f" containerName="ceilometer-notification-agent" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.132730 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5495b02-2ce5-4d67-8dbf-869d08e8891f" containerName="ceilometer-central-agent" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.132740 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5495b02-2ce5-4d67-8dbf-869d08e8891f" containerName="proxy-httpd" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.135486 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.139688 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.140557 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.141980 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.147075 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.167417 4934 scope.go:117] "RemoveContainer" containerID="e03df11cd32b7ae16694526d9e9ba028080674e0673db23fedec649bbf3e5289" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.188525 4934 scope.go:117] "RemoveContainer" containerID="3d1e128e9cd40b07a6e5b793c66fedbe9bcd9b29d9d42be19604636c8786b248" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.313577 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.313626 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f370334-c935-4266-843c-16854e04b27c-run-httpd\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.313688 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-config-data\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.313711 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.313795 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f370334-c935-4266-843c-16854e04b27c-log-httpd\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.313860 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-scripts\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.313924 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82m2g\" (UniqueName: \"kubernetes.io/projected/3f370334-c935-4266-843c-16854e04b27c-kube-api-access-82m2g\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.313943 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.416199 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f370334-c935-4266-843c-16854e04b27c-log-httpd\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.416523 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-scripts\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.416579 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82m2g\" (UniqueName: \"kubernetes.io/projected/3f370334-c935-4266-843c-16854e04b27c-kube-api-access-82m2g\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.416604 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.416655 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.416682 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f370334-c935-4266-843c-16854e04b27c-run-httpd\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.416716 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-config-data\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.416736 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.417473 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f370334-c935-4266-843c-16854e04b27c-run-httpd\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.417907 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f370334-c935-4266-843c-16854e04b27c-log-httpd\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.423223 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-scripts\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.423563 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.427856 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.429479 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-config-data\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.445719 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82m2g\" (UniqueName: \"kubernetes.io/projected/3f370334-c935-4266-843c-16854e04b27c-kube-api-access-82m2g\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.447480 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " pod="openstack/ceilometer-0" Dec 27 08:08:36 crc kubenswrapper[4934]: I1227 08:08:36.471662 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:08:37 crc kubenswrapper[4934]: I1227 08:08:37.067378 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g8b74" event={"ID":"784ae8a1-3984-4b1d-830d-203e05f150df","Type":"ContainerStarted","Data":"957303f2897e8ca3301ff38b9e0e845a932e6554a706fe8bb818c8e616b1637b"} Dec 27 08:08:37 crc kubenswrapper[4934]: I1227 08:08:37.090838 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-g8b74" podStartSLOduration=3.336091334 podStartE2EDuration="6.090819938s" podCreationTimestamp="2025-12-27 08:08:31 +0000 UTC" firstStartedPulling="2025-12-27 08:08:33.999439051 +0000 UTC m=+1574.819879655" lastFinishedPulling="2025-12-27 08:08:36.754167665 +0000 UTC m=+1577.574608259" observedRunningTime="2025-12-27 08:08:37.080855571 +0000 UTC m=+1577.901296155" watchObservedRunningTime="2025-12-27 08:08:37.090819938 +0000 UTC m=+1577.911260532" Dec 27 08:08:37 crc kubenswrapper[4934]: I1227 08:08:37.150160 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:08:37 crc kubenswrapper[4934]: I1227 08:08:37.454666 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-5qllw"] Dec 27 08:08:37 crc kubenswrapper[4934]: I1227 08:08:37.465438 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-5qllw"] Dec 27 08:08:37 crc kubenswrapper[4934]: I1227 08:08:37.534037 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61f5132a-303d-4f76-ae58-c57d82a59fd4" path="/var/lib/kubelet/pods/61f5132a-303d-4f76-ae58-c57d82a59fd4/volumes" Dec 27 08:08:37 crc kubenswrapper[4934]: I1227 08:08:37.534814 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5495b02-2ce5-4d67-8dbf-869d08e8891f" path="/var/lib/kubelet/pods/f5495b02-2ce5-4d67-8dbf-869d08e8891f/volumes" Dec 27 08:08:37 crc kubenswrapper[4934]: I1227 08:08:37.558145 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-njzm7"] Dec 27 08:08:37 crc kubenswrapper[4934]: I1227 08:08:37.561052 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-njzm7" Dec 27 08:08:37 crc kubenswrapper[4934]: I1227 08:08:37.585420 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-njzm7"] Dec 27 08:08:37 crc kubenswrapper[4934]: I1227 08:08:37.648706 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45927aae-ef6c-475a-b48f-59cd34ff1278-combined-ca-bundle\") pod \"heat-db-sync-njzm7\" (UID: \"45927aae-ef6c-475a-b48f-59cd34ff1278\") " pod="openstack/heat-db-sync-njzm7" Dec 27 08:08:37 crc kubenswrapper[4934]: I1227 08:08:37.649076 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlvw8\" (UniqueName: \"kubernetes.io/projected/45927aae-ef6c-475a-b48f-59cd34ff1278-kube-api-access-tlvw8\") pod \"heat-db-sync-njzm7\" (UID: \"45927aae-ef6c-475a-b48f-59cd34ff1278\") " pod="openstack/heat-db-sync-njzm7" Dec 27 08:08:37 crc kubenswrapper[4934]: I1227 08:08:37.649208 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45927aae-ef6c-475a-b48f-59cd34ff1278-config-data\") pod \"heat-db-sync-njzm7\" (UID: \"45927aae-ef6c-475a-b48f-59cd34ff1278\") " pod="openstack/heat-db-sync-njzm7" Dec 27 08:08:37 crc kubenswrapper[4934]: I1227 08:08:37.750986 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45927aae-ef6c-475a-b48f-59cd34ff1278-config-data\") pod \"heat-db-sync-njzm7\" (UID: \"45927aae-ef6c-475a-b48f-59cd34ff1278\") " pod="openstack/heat-db-sync-njzm7" Dec 27 08:08:37 crc kubenswrapper[4934]: I1227 08:08:37.751140 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45927aae-ef6c-475a-b48f-59cd34ff1278-combined-ca-bundle\") pod \"heat-db-sync-njzm7\" (UID: \"45927aae-ef6c-475a-b48f-59cd34ff1278\") " pod="openstack/heat-db-sync-njzm7" Dec 27 08:08:37 crc kubenswrapper[4934]: I1227 08:08:37.751228 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlvw8\" (UniqueName: \"kubernetes.io/projected/45927aae-ef6c-475a-b48f-59cd34ff1278-kube-api-access-tlvw8\") pod \"heat-db-sync-njzm7\" (UID: \"45927aae-ef6c-475a-b48f-59cd34ff1278\") " pod="openstack/heat-db-sync-njzm7" Dec 27 08:08:37 crc kubenswrapper[4934]: I1227 08:08:37.758380 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45927aae-ef6c-475a-b48f-59cd34ff1278-config-data\") pod \"heat-db-sync-njzm7\" (UID: \"45927aae-ef6c-475a-b48f-59cd34ff1278\") " pod="openstack/heat-db-sync-njzm7" Dec 27 08:08:37 crc kubenswrapper[4934]: I1227 08:08:37.761060 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45927aae-ef6c-475a-b48f-59cd34ff1278-combined-ca-bundle\") pod \"heat-db-sync-njzm7\" (UID: \"45927aae-ef6c-475a-b48f-59cd34ff1278\") " pod="openstack/heat-db-sync-njzm7" Dec 27 08:08:37 crc kubenswrapper[4934]: I1227 08:08:37.780475 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlvw8\" (UniqueName: \"kubernetes.io/projected/45927aae-ef6c-475a-b48f-59cd34ff1278-kube-api-access-tlvw8\") pod \"heat-db-sync-njzm7\" (UID: \"45927aae-ef6c-475a-b48f-59cd34ff1278\") " pod="openstack/heat-db-sync-njzm7" Dec 27 08:08:37 crc kubenswrapper[4934]: I1227 08:08:37.885575 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-njzm7" Dec 27 08:08:38 crc kubenswrapper[4934]: I1227 08:08:38.079320 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f370334-c935-4266-843c-16854e04b27c","Type":"ContainerStarted","Data":"ac2512801412add05bc89622beca704302346f56101e624aa6c41ecc7c4be65d"} Dec 27 08:08:38 crc kubenswrapper[4934]: I1227 08:08:38.079698 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f370334-c935-4266-843c-16854e04b27c","Type":"ContainerStarted","Data":"838d962663f8bc34fbb4f3e3aea4a55d4dc4a15ef4440ab114431728d687eac2"} Dec 27 08:08:38 crc kubenswrapper[4934]: I1227 08:08:38.393048 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-njzm7"] Dec 27 08:08:39 crc kubenswrapper[4934]: I1227 08:08:39.099021 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-njzm7" event={"ID":"45927aae-ef6c-475a-b48f-59cd34ff1278","Type":"ContainerStarted","Data":"367a569da7ac5ff50eaf33b34df280128fff893009c17780dde69a5c99dca229"} Dec 27 08:08:39 crc kubenswrapper[4934]: I1227 08:08:39.102707 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f370334-c935-4266-843c-16854e04b27c","Type":"ContainerStarted","Data":"0f232ba798b2584fe144d6a2393ab39ab1fd1614e734a927949f7b09506fbe6d"} Dec 27 08:08:39 crc kubenswrapper[4934]: I1227 08:08:39.590151 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-2"] Dec 27 08:08:40 crc kubenswrapper[4934]: I1227 08:08:40.118177 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f370334-c935-4266-843c-16854e04b27c","Type":"ContainerStarted","Data":"c91492ea0d969b5d2b1466424daeca7b7111228fb310f0108c239622f3f87b74"} Dec 27 08:08:40 crc kubenswrapper[4934]: I1227 08:08:40.468916 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:08:40 crc kubenswrapper[4934]: E1227 08:08:40.469528 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:08:41 crc kubenswrapper[4934]: I1227 08:08:41.034583 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 27 08:08:41 crc kubenswrapper[4934]: I1227 08:08:41.418690 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 27 08:08:42 crc kubenswrapper[4934]: I1227 08:08:42.043571 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:08:42 crc kubenswrapper[4934]: I1227 08:08:42.168232 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f370334-c935-4266-843c-16854e04b27c","Type":"ContainerStarted","Data":"a5dca8402d5577ae1f7d47f1ab29cf325408470b718d76b7839e05fddf394456"} Dec 27 08:08:42 crc kubenswrapper[4934]: I1227 08:08:42.168582 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 27 08:08:42 crc kubenswrapper[4934]: I1227 08:08:42.221913 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.335765316 podStartE2EDuration="6.22189428s" podCreationTimestamp="2025-12-27 08:08:36 +0000 UTC" firstStartedPulling="2025-12-27 08:08:37.159386479 +0000 UTC m=+1577.979827073" lastFinishedPulling="2025-12-27 08:08:41.045515443 +0000 UTC m=+1581.865956037" observedRunningTime="2025-12-27 08:08:42.210658791 +0000 UTC m=+1583.031099385" watchObservedRunningTime="2025-12-27 08:08:42.22189428 +0000 UTC m=+1583.042334874" Dec 27 08:08:42 crc kubenswrapper[4934]: I1227 08:08:42.289961 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-g8b74" Dec 27 08:08:42 crc kubenswrapper[4934]: I1227 08:08:42.290000 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-g8b74" Dec 27 08:08:42 crc kubenswrapper[4934]: I1227 08:08:42.368395 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-g8b74" Dec 27 08:08:43 crc kubenswrapper[4934]: I1227 08:08:43.180265 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3f370334-c935-4266-843c-16854e04b27c" containerName="ceilometer-central-agent" containerID="cri-o://ac2512801412add05bc89622beca704302346f56101e624aa6c41ecc7c4be65d" gracePeriod=30 Dec 27 08:08:43 crc kubenswrapper[4934]: I1227 08:08:43.180513 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3f370334-c935-4266-843c-16854e04b27c" containerName="ceilometer-notification-agent" containerID="cri-o://0f232ba798b2584fe144d6a2393ab39ab1fd1614e734a927949f7b09506fbe6d" gracePeriod=30 Dec 27 08:08:43 crc kubenswrapper[4934]: I1227 08:08:43.180518 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3f370334-c935-4266-843c-16854e04b27c" containerName="sg-core" containerID="cri-o://c91492ea0d969b5d2b1466424daeca7b7111228fb310f0108c239622f3f87b74" gracePeriod=30 Dec 27 08:08:43 crc kubenswrapper[4934]: I1227 08:08:43.181205 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3f370334-c935-4266-843c-16854e04b27c" containerName="proxy-httpd" containerID="cri-o://a5dca8402d5577ae1f7d47f1ab29cf325408470b718d76b7839e05fddf394456" gracePeriod=30 Dec 27 08:08:43 crc kubenswrapper[4934]: I1227 08:08:43.260606 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-g8b74" Dec 27 08:08:43 crc kubenswrapper[4934]: I1227 08:08:43.337379 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g8b74"] Dec 27 08:08:44 crc kubenswrapper[4934]: I1227 08:08:44.193525 4934 generic.go:334] "Generic (PLEG): container finished" podID="3f370334-c935-4266-843c-16854e04b27c" containerID="a5dca8402d5577ae1f7d47f1ab29cf325408470b718d76b7839e05fddf394456" exitCode=0 Dec 27 08:08:44 crc kubenswrapper[4934]: I1227 08:08:44.193790 4934 generic.go:334] "Generic (PLEG): container finished" podID="3f370334-c935-4266-843c-16854e04b27c" containerID="c91492ea0d969b5d2b1466424daeca7b7111228fb310f0108c239622f3f87b74" exitCode=2 Dec 27 08:08:44 crc kubenswrapper[4934]: I1227 08:08:44.193801 4934 generic.go:334] "Generic (PLEG): container finished" podID="3f370334-c935-4266-843c-16854e04b27c" containerID="0f232ba798b2584fe144d6a2393ab39ab1fd1614e734a927949f7b09506fbe6d" exitCode=0 Dec 27 08:08:44 crc kubenswrapper[4934]: I1227 08:08:44.193737 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f370334-c935-4266-843c-16854e04b27c","Type":"ContainerDied","Data":"a5dca8402d5577ae1f7d47f1ab29cf325408470b718d76b7839e05fddf394456"} Dec 27 08:08:44 crc kubenswrapper[4934]: I1227 08:08:44.194819 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f370334-c935-4266-843c-16854e04b27c","Type":"ContainerDied","Data":"c91492ea0d969b5d2b1466424daeca7b7111228fb310f0108c239622f3f87b74"} Dec 27 08:08:44 crc kubenswrapper[4934]: I1227 08:08:44.194886 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f370334-c935-4266-843c-16854e04b27c","Type":"ContainerDied","Data":"0f232ba798b2584fe144d6a2393ab39ab1fd1614e734a927949f7b09506fbe6d"} Dec 27 08:08:45 crc kubenswrapper[4934]: I1227 08:08:45.203743 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-g8b74" podUID="784ae8a1-3984-4b1d-830d-203e05f150df" containerName="registry-server" containerID="cri-o://957303f2897e8ca3301ff38b9e0e845a932e6554a706fe8bb818c8e616b1637b" gracePeriod=2 Dec 27 08:08:45 crc kubenswrapper[4934]: I1227 08:08:45.230527 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="76158d63-2099-4418-82a2-09af0432c1a3" containerName="rabbitmq" containerID="cri-o://d137c3d7f1fb61a5492ae4e87102e9b26b5caceeac796ed7e5db20c3505e981d" gracePeriod=604796 Dec 27 08:08:45 crc kubenswrapper[4934]: I1227 08:08:45.313720 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-2" podUID="df7d6b06-cb87-4635-8aca-1ab00a564dd7" containerName="rabbitmq" containerID="cri-o://7572ade40e346d8651cbf1ce8fabe437ad6fbf893725ab3cc22288f68c146c30" gracePeriod=604795 Dec 27 08:08:45 crc kubenswrapper[4934]: I1227 08:08:45.842242 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:08:45 crc kubenswrapper[4934]: I1227 08:08:45.949890 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g8b74" Dec 27 08:08:45 crc kubenswrapper[4934]: I1227 08:08:45.980721 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-sg-core-conf-yaml\") pod \"3f370334-c935-4266-843c-16854e04b27c\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " Dec 27 08:08:45 crc kubenswrapper[4934]: I1227 08:08:45.981168 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f370334-c935-4266-843c-16854e04b27c-log-httpd\") pod \"3f370334-c935-4266-843c-16854e04b27c\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " Dec 27 08:08:45 crc kubenswrapper[4934]: I1227 08:08:45.981245 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-ceilometer-tls-certs\") pod \"3f370334-c935-4266-843c-16854e04b27c\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " Dec 27 08:08:45 crc kubenswrapper[4934]: I1227 08:08:45.981361 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-combined-ca-bundle\") pod \"3f370334-c935-4266-843c-16854e04b27c\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " Dec 27 08:08:45 crc kubenswrapper[4934]: I1227 08:08:45.981472 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-scripts\") pod \"3f370334-c935-4266-843c-16854e04b27c\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " Dec 27 08:08:45 crc kubenswrapper[4934]: I1227 08:08:45.981557 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-config-data\") pod \"3f370334-c935-4266-843c-16854e04b27c\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " Dec 27 08:08:45 crc kubenswrapper[4934]: I1227 08:08:45.981712 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82m2g\" (UniqueName: \"kubernetes.io/projected/3f370334-c935-4266-843c-16854e04b27c-kube-api-access-82m2g\") pod \"3f370334-c935-4266-843c-16854e04b27c\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " Dec 27 08:08:45 crc kubenswrapper[4934]: I1227 08:08:45.981788 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f370334-c935-4266-843c-16854e04b27c-run-httpd\") pod \"3f370334-c935-4266-843c-16854e04b27c\" (UID: \"3f370334-c935-4266-843c-16854e04b27c\") " Dec 27 08:08:45 crc kubenswrapper[4934]: I1227 08:08:45.981992 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f370334-c935-4266-843c-16854e04b27c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3f370334-c935-4266-843c-16854e04b27c" (UID: "3f370334-c935-4266-843c-16854e04b27c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:08:45 crc kubenswrapper[4934]: I1227 08:08:45.982426 4934 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f370334-c935-4266-843c-16854e04b27c-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:45 crc kubenswrapper[4934]: I1227 08:08:45.982830 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f370334-c935-4266-843c-16854e04b27c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3f370334-c935-4266-843c-16854e04b27c" (UID: "3f370334-c935-4266-843c-16854e04b27c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:08:45 crc kubenswrapper[4934]: I1227 08:08:45.987672 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-scripts" (OuterVolumeSpecName: "scripts") pod "3f370334-c935-4266-843c-16854e04b27c" (UID: "3f370334-c935-4266-843c-16854e04b27c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:08:45 crc kubenswrapper[4934]: I1227 08:08:45.989224 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f370334-c935-4266-843c-16854e04b27c-kube-api-access-82m2g" (OuterVolumeSpecName: "kube-api-access-82m2g") pod "3f370334-c935-4266-843c-16854e04b27c" (UID: "3f370334-c935-4266-843c-16854e04b27c"). InnerVolumeSpecName "kube-api-access-82m2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.017969 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3f370334-c935-4266-843c-16854e04b27c" (UID: "3f370334-c935-4266-843c-16854e04b27c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.066569 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "3f370334-c935-4266-843c-16854e04b27c" (UID: "3f370334-c935-4266-843c-16854e04b27c"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.084216 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxm89\" (UniqueName: \"kubernetes.io/projected/784ae8a1-3984-4b1d-830d-203e05f150df-kube-api-access-rxm89\") pod \"784ae8a1-3984-4b1d-830d-203e05f150df\" (UID: \"784ae8a1-3984-4b1d-830d-203e05f150df\") " Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.084393 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/784ae8a1-3984-4b1d-830d-203e05f150df-catalog-content\") pod \"784ae8a1-3984-4b1d-830d-203e05f150df\" (UID: \"784ae8a1-3984-4b1d-830d-203e05f150df\") " Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.084445 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/784ae8a1-3984-4b1d-830d-203e05f150df-utilities\") pod \"784ae8a1-3984-4b1d-830d-203e05f150df\" (UID: \"784ae8a1-3984-4b1d-830d-203e05f150df\") " Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.085497 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/784ae8a1-3984-4b1d-830d-203e05f150df-utilities" (OuterVolumeSpecName: "utilities") pod "784ae8a1-3984-4b1d-830d-203e05f150df" (UID: "784ae8a1-3984-4b1d-830d-203e05f150df"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.088717 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f370334-c935-4266-843c-16854e04b27c" (UID: "3f370334-c935-4266-843c-16854e04b27c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.089159 4934 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.089192 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.089325 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82m2g\" (UniqueName: \"kubernetes.io/projected/3f370334-c935-4266-843c-16854e04b27c-kube-api-access-82m2g\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.089338 4934 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3f370334-c935-4266-843c-16854e04b27c-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.089348 4934 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.089357 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/784ae8a1-3984-4b1d-830d-203e05f150df-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.094871 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/784ae8a1-3984-4b1d-830d-203e05f150df-kube-api-access-rxm89" (OuterVolumeSpecName: "kube-api-access-rxm89") pod "784ae8a1-3984-4b1d-830d-203e05f150df" (UID: "784ae8a1-3984-4b1d-830d-203e05f150df"). InnerVolumeSpecName "kube-api-access-rxm89". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.104761 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-config-data" (OuterVolumeSpecName: "config-data") pod "3f370334-c935-4266-843c-16854e04b27c" (UID: "3f370334-c935-4266-843c-16854e04b27c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.107329 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/784ae8a1-3984-4b1d-830d-203e05f150df-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "784ae8a1-3984-4b1d-830d-203e05f150df" (UID: "784ae8a1-3984-4b1d-830d-203e05f150df"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.191349 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.191605 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f370334-c935-4266-843c-16854e04b27c-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.191616 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxm89\" (UniqueName: \"kubernetes.io/projected/784ae8a1-3984-4b1d-830d-203e05f150df-kube-api-access-rxm89\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.191628 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/784ae8a1-3984-4b1d-830d-203e05f150df-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.217617 4934 generic.go:334] "Generic (PLEG): container finished" podID="784ae8a1-3984-4b1d-830d-203e05f150df" containerID="957303f2897e8ca3301ff38b9e0e845a932e6554a706fe8bb818c8e616b1637b" exitCode=0 Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.217684 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g8b74" event={"ID":"784ae8a1-3984-4b1d-830d-203e05f150df","Type":"ContainerDied","Data":"957303f2897e8ca3301ff38b9e0e845a932e6554a706fe8bb818c8e616b1637b"} Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.217703 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g8b74" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.217736 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g8b74" event={"ID":"784ae8a1-3984-4b1d-830d-203e05f150df","Type":"ContainerDied","Data":"942a7040e8f92f86bd08615728c5008b8dbee5e9193d2253b6faebe3499fca28"} Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.217755 4934 scope.go:117] "RemoveContainer" containerID="957303f2897e8ca3301ff38b9e0e845a932e6554a706fe8bb818c8e616b1637b" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.225639 4934 generic.go:334] "Generic (PLEG): container finished" podID="3f370334-c935-4266-843c-16854e04b27c" containerID="ac2512801412add05bc89622beca704302346f56101e624aa6c41ecc7c4be65d" exitCode=0 Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.225667 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f370334-c935-4266-843c-16854e04b27c","Type":"ContainerDied","Data":"ac2512801412add05bc89622beca704302346f56101e624aa6c41ecc7c4be65d"} Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.225692 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3f370334-c935-4266-843c-16854e04b27c","Type":"ContainerDied","Data":"838d962663f8bc34fbb4f3e3aea4a55d4dc4a15ef4440ab114431728d687eac2"} Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.225745 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.308489 4934 scope.go:117] "RemoveContainer" containerID="b21f5414f614273e6d18af71037c49ddc262adbf8a2f8ac8da5ef144ee2d597d" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.327363 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g8b74"] Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.348786 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-g8b74"] Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.355248 4934 scope.go:117] "RemoveContainer" containerID="2c92046d337fa00ce7799cb0163cc967278772f32ddfc48be0b7fa748b271665" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.370130 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.405891 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.423594 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:08:46 crc kubenswrapper[4934]: E1227 08:08:46.424214 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f370334-c935-4266-843c-16854e04b27c" containerName="proxy-httpd" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.424233 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f370334-c935-4266-843c-16854e04b27c" containerName="proxy-httpd" Dec 27 08:08:46 crc kubenswrapper[4934]: E1227 08:08:46.424253 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="784ae8a1-3984-4b1d-830d-203e05f150df" containerName="extract-content" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.424259 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="784ae8a1-3984-4b1d-830d-203e05f150df" containerName="extract-content" Dec 27 08:08:46 crc kubenswrapper[4934]: E1227 08:08:46.424278 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f370334-c935-4266-843c-16854e04b27c" containerName="ceilometer-notification-agent" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.424285 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f370334-c935-4266-843c-16854e04b27c" containerName="ceilometer-notification-agent" Dec 27 08:08:46 crc kubenswrapper[4934]: E1227 08:08:46.424299 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="784ae8a1-3984-4b1d-830d-203e05f150df" containerName="extract-utilities" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.424306 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="784ae8a1-3984-4b1d-830d-203e05f150df" containerName="extract-utilities" Dec 27 08:08:46 crc kubenswrapper[4934]: E1227 08:08:46.424329 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f370334-c935-4266-843c-16854e04b27c" containerName="sg-core" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.424335 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f370334-c935-4266-843c-16854e04b27c" containerName="sg-core" Dec 27 08:08:46 crc kubenswrapper[4934]: E1227 08:08:46.424351 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="784ae8a1-3984-4b1d-830d-203e05f150df" containerName="registry-server" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.424357 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="784ae8a1-3984-4b1d-830d-203e05f150df" containerName="registry-server" Dec 27 08:08:46 crc kubenswrapper[4934]: E1227 08:08:46.424372 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f370334-c935-4266-843c-16854e04b27c" containerName="ceilometer-central-agent" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.424377 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f370334-c935-4266-843c-16854e04b27c" containerName="ceilometer-central-agent" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.424600 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f370334-c935-4266-843c-16854e04b27c" containerName="sg-core" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.424612 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="784ae8a1-3984-4b1d-830d-203e05f150df" containerName="registry-server" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.424625 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f370334-c935-4266-843c-16854e04b27c" containerName="ceilometer-notification-agent" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.424644 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f370334-c935-4266-843c-16854e04b27c" containerName="ceilometer-central-agent" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.424655 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f370334-c935-4266-843c-16854e04b27c" containerName="proxy-httpd" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.426715 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.429762 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.430592 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.434406 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.437772 4934 scope.go:117] "RemoveContainer" containerID="957303f2897e8ca3301ff38b9e0e845a932e6554a706fe8bb818c8e616b1637b" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.439187 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:08:46 crc kubenswrapper[4934]: E1227 08:08:46.439735 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"957303f2897e8ca3301ff38b9e0e845a932e6554a706fe8bb818c8e616b1637b\": container with ID starting with 957303f2897e8ca3301ff38b9e0e845a932e6554a706fe8bb818c8e616b1637b not found: ID does not exist" containerID="957303f2897e8ca3301ff38b9e0e845a932e6554a706fe8bb818c8e616b1637b" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.439817 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"957303f2897e8ca3301ff38b9e0e845a932e6554a706fe8bb818c8e616b1637b"} err="failed to get container status \"957303f2897e8ca3301ff38b9e0e845a932e6554a706fe8bb818c8e616b1637b\": rpc error: code = NotFound desc = could not find container \"957303f2897e8ca3301ff38b9e0e845a932e6554a706fe8bb818c8e616b1637b\": container with ID starting with 957303f2897e8ca3301ff38b9e0e845a932e6554a706fe8bb818c8e616b1637b not found: ID does not exist" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.439899 4934 scope.go:117] "RemoveContainer" containerID="b21f5414f614273e6d18af71037c49ddc262adbf8a2f8ac8da5ef144ee2d597d" Dec 27 08:08:46 crc kubenswrapper[4934]: E1227 08:08:46.440416 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b21f5414f614273e6d18af71037c49ddc262adbf8a2f8ac8da5ef144ee2d597d\": container with ID starting with b21f5414f614273e6d18af71037c49ddc262adbf8a2f8ac8da5ef144ee2d597d not found: ID does not exist" containerID="b21f5414f614273e6d18af71037c49ddc262adbf8a2f8ac8da5ef144ee2d597d" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.440502 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b21f5414f614273e6d18af71037c49ddc262adbf8a2f8ac8da5ef144ee2d597d"} err="failed to get container status \"b21f5414f614273e6d18af71037c49ddc262adbf8a2f8ac8da5ef144ee2d597d\": rpc error: code = NotFound desc = could not find container \"b21f5414f614273e6d18af71037c49ddc262adbf8a2f8ac8da5ef144ee2d597d\": container with ID starting with b21f5414f614273e6d18af71037c49ddc262adbf8a2f8ac8da5ef144ee2d597d not found: ID does not exist" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.440572 4934 scope.go:117] "RemoveContainer" containerID="2c92046d337fa00ce7799cb0163cc967278772f32ddfc48be0b7fa748b271665" Dec 27 08:08:46 crc kubenswrapper[4934]: E1227 08:08:46.440888 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c92046d337fa00ce7799cb0163cc967278772f32ddfc48be0b7fa748b271665\": container with ID starting with 2c92046d337fa00ce7799cb0163cc967278772f32ddfc48be0b7fa748b271665 not found: ID does not exist" containerID="2c92046d337fa00ce7799cb0163cc967278772f32ddfc48be0b7fa748b271665" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.440985 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c92046d337fa00ce7799cb0163cc967278772f32ddfc48be0b7fa748b271665"} err="failed to get container status \"2c92046d337fa00ce7799cb0163cc967278772f32ddfc48be0b7fa748b271665\": rpc error: code = NotFound desc = could not find container \"2c92046d337fa00ce7799cb0163cc967278772f32ddfc48be0b7fa748b271665\": container with ID starting with 2c92046d337fa00ce7799cb0163cc967278772f32ddfc48be0b7fa748b271665 not found: ID does not exist" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.441070 4934 scope.go:117] "RemoveContainer" containerID="a5dca8402d5577ae1f7d47f1ab29cf325408470b718d76b7839e05fddf394456" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.502718 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.502806 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzb6v\" (UniqueName: \"kubernetes.io/projected/12e88dda-31ca-4b5b-93c3-7eced0cacda7-kube-api-access-qzb6v\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.502836 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.502866 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-config-data\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.502970 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12e88dda-31ca-4b5b-93c3-7eced0cacda7-log-httpd\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.502997 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-scripts\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.503104 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.504239 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12e88dda-31ca-4b5b-93c3-7eced0cacda7-run-httpd\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.508739 4934 scope.go:117] "RemoveContainer" containerID="c91492ea0d969b5d2b1466424daeca7b7111228fb310f0108c239622f3f87b74" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.535377 4934 scope.go:117] "RemoveContainer" containerID="0f232ba798b2584fe144d6a2393ab39ab1fd1614e734a927949f7b09506fbe6d" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.556600 4934 scope.go:117] "RemoveContainer" containerID="ac2512801412add05bc89622beca704302346f56101e624aa6c41ecc7c4be65d" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.583703 4934 scope.go:117] "RemoveContainer" containerID="a5dca8402d5577ae1f7d47f1ab29cf325408470b718d76b7839e05fddf394456" Dec 27 08:08:46 crc kubenswrapper[4934]: E1227 08:08:46.584233 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5dca8402d5577ae1f7d47f1ab29cf325408470b718d76b7839e05fddf394456\": container with ID starting with a5dca8402d5577ae1f7d47f1ab29cf325408470b718d76b7839e05fddf394456 not found: ID does not exist" containerID="a5dca8402d5577ae1f7d47f1ab29cf325408470b718d76b7839e05fddf394456" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.584276 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5dca8402d5577ae1f7d47f1ab29cf325408470b718d76b7839e05fddf394456"} err="failed to get container status \"a5dca8402d5577ae1f7d47f1ab29cf325408470b718d76b7839e05fddf394456\": rpc error: code = NotFound desc = could not find container \"a5dca8402d5577ae1f7d47f1ab29cf325408470b718d76b7839e05fddf394456\": container with ID starting with a5dca8402d5577ae1f7d47f1ab29cf325408470b718d76b7839e05fddf394456 not found: ID does not exist" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.584314 4934 scope.go:117] "RemoveContainer" containerID="c91492ea0d969b5d2b1466424daeca7b7111228fb310f0108c239622f3f87b74" Dec 27 08:08:46 crc kubenswrapper[4934]: E1227 08:08:46.584862 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c91492ea0d969b5d2b1466424daeca7b7111228fb310f0108c239622f3f87b74\": container with ID starting with c91492ea0d969b5d2b1466424daeca7b7111228fb310f0108c239622f3f87b74 not found: ID does not exist" containerID="c91492ea0d969b5d2b1466424daeca7b7111228fb310f0108c239622f3f87b74" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.584882 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c91492ea0d969b5d2b1466424daeca7b7111228fb310f0108c239622f3f87b74"} err="failed to get container status \"c91492ea0d969b5d2b1466424daeca7b7111228fb310f0108c239622f3f87b74\": rpc error: code = NotFound desc = could not find container \"c91492ea0d969b5d2b1466424daeca7b7111228fb310f0108c239622f3f87b74\": container with ID starting with c91492ea0d969b5d2b1466424daeca7b7111228fb310f0108c239622f3f87b74 not found: ID does not exist" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.584895 4934 scope.go:117] "RemoveContainer" containerID="0f232ba798b2584fe144d6a2393ab39ab1fd1614e734a927949f7b09506fbe6d" Dec 27 08:08:46 crc kubenswrapper[4934]: E1227 08:08:46.585227 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f232ba798b2584fe144d6a2393ab39ab1fd1614e734a927949f7b09506fbe6d\": container with ID starting with 0f232ba798b2584fe144d6a2393ab39ab1fd1614e734a927949f7b09506fbe6d not found: ID does not exist" containerID="0f232ba798b2584fe144d6a2393ab39ab1fd1614e734a927949f7b09506fbe6d" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.585245 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f232ba798b2584fe144d6a2393ab39ab1fd1614e734a927949f7b09506fbe6d"} err="failed to get container status \"0f232ba798b2584fe144d6a2393ab39ab1fd1614e734a927949f7b09506fbe6d\": rpc error: code = NotFound desc = could not find container \"0f232ba798b2584fe144d6a2393ab39ab1fd1614e734a927949f7b09506fbe6d\": container with ID starting with 0f232ba798b2584fe144d6a2393ab39ab1fd1614e734a927949f7b09506fbe6d not found: ID does not exist" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.585257 4934 scope.go:117] "RemoveContainer" containerID="ac2512801412add05bc89622beca704302346f56101e624aa6c41ecc7c4be65d" Dec 27 08:08:46 crc kubenswrapper[4934]: E1227 08:08:46.585490 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac2512801412add05bc89622beca704302346f56101e624aa6c41ecc7c4be65d\": container with ID starting with ac2512801412add05bc89622beca704302346f56101e624aa6c41ecc7c4be65d not found: ID does not exist" containerID="ac2512801412add05bc89622beca704302346f56101e624aa6c41ecc7c4be65d" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.585515 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac2512801412add05bc89622beca704302346f56101e624aa6c41ecc7c4be65d"} err="failed to get container status \"ac2512801412add05bc89622beca704302346f56101e624aa6c41ecc7c4be65d\": rpc error: code = NotFound desc = could not find container \"ac2512801412add05bc89622beca704302346f56101e624aa6c41ecc7c4be65d\": container with ID starting with ac2512801412add05bc89622beca704302346f56101e624aa6c41ecc7c4be65d not found: ID does not exist" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.606328 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.606396 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzb6v\" (UniqueName: \"kubernetes.io/projected/12e88dda-31ca-4b5b-93c3-7eced0cacda7-kube-api-access-qzb6v\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.606419 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.606448 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-config-data\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.606547 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12e88dda-31ca-4b5b-93c3-7eced0cacda7-log-httpd\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.606893 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-scripts\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.607025 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.607074 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12e88dda-31ca-4b5b-93c3-7eced0cacda7-run-httpd\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.607070 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12e88dda-31ca-4b5b-93c3-7eced0cacda7-log-httpd\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.607657 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12e88dda-31ca-4b5b-93c3-7eced0cacda7-run-httpd\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.611958 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.612525 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.613076 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.613886 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-scripts\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.615338 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-config-data\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.622477 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzb6v\" (UniqueName: \"kubernetes.io/projected/12e88dda-31ca-4b5b-93c3-7eced0cacda7-kube-api-access-qzb6v\") pod \"ceilometer-0\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " pod="openstack/ceilometer-0" Dec 27 08:08:46 crc kubenswrapper[4934]: I1227 08:08:46.821646 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 08:08:47 crc kubenswrapper[4934]: I1227 08:08:47.499219 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f370334-c935-4266-843c-16854e04b27c" path="/var/lib/kubelet/pods/3f370334-c935-4266-843c-16854e04b27c/volumes" Dec 27 08:08:47 crc kubenswrapper[4934]: I1227 08:08:47.499924 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="784ae8a1-3984-4b1d-830d-203e05f150df" path="/var/lib/kubelet/pods/784ae8a1-3984-4b1d-830d-203e05f150df/volumes" Dec 27 08:08:48 crc kubenswrapper[4934]: I1227 08:08:48.850721 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="df7d6b06-cb87-4635-8aca-1ab00a564dd7" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.131:5671: connect: connection refused" Dec 27 08:08:49 crc kubenswrapper[4934]: I1227 08:08:49.044364 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="76158d63-2099-4418-82a2-09af0432c1a3" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.132:5671: connect: connection refused" Dec 27 08:08:52 crc kubenswrapper[4934]: I1227 08:08:52.321126 4934 generic.go:334] "Generic (PLEG): container finished" podID="df7d6b06-cb87-4635-8aca-1ab00a564dd7" containerID="7572ade40e346d8651cbf1ce8fabe437ad6fbf893725ab3cc22288f68c146c30" exitCode=0 Dec 27 08:08:52 crc kubenswrapper[4934]: I1227 08:08:52.321281 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"df7d6b06-cb87-4635-8aca-1ab00a564dd7","Type":"ContainerDied","Data":"7572ade40e346d8651cbf1ce8fabe437ad6fbf893725ab3cc22288f68c146c30"} Dec 27 08:08:52 crc kubenswrapper[4934]: I1227 08:08:52.327299 4934 generic.go:334] "Generic (PLEG): container finished" podID="76158d63-2099-4418-82a2-09af0432c1a3" containerID="d137c3d7f1fb61a5492ae4e87102e9b26b5caceeac796ed7e5db20c3505e981d" exitCode=0 Dec 27 08:08:52 crc kubenswrapper[4934]: I1227 08:08:52.327333 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"76158d63-2099-4418-82a2-09af0432c1a3","Type":"ContainerDied","Data":"d137c3d7f1fb61a5492ae4e87102e9b26b5caceeac796ed7e5db20c3505e981d"} Dec 27 08:08:52 crc kubenswrapper[4934]: I1227 08:08:52.468531 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:08:52 crc kubenswrapper[4934]: E1227 08:08:52.468823 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:08:54 crc kubenswrapper[4934]: I1227 08:08:54.816455 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-8g7tc"] Dec 27 08:08:54 crc kubenswrapper[4934]: I1227 08:08:54.836941 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:54 crc kubenswrapper[4934]: I1227 08:08:54.841422 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-8g7tc"] Dec 27 08:08:54 crc kubenswrapper[4934]: I1227 08:08:54.855756 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 27 08:08:54 crc kubenswrapper[4934]: I1227 08:08:54.924866 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-ovsdbserver-sb\") pod \"dnsmasq-dns-5b75489c6f-8g7tc\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:54 crc kubenswrapper[4934]: I1227 08:08:54.924957 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-dns-swift-storage-0\") pod \"dnsmasq-dns-5b75489c6f-8g7tc\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:54 crc kubenswrapper[4934]: I1227 08:08:54.925113 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-dns-svc\") pod \"dnsmasq-dns-5b75489c6f-8g7tc\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:54 crc kubenswrapper[4934]: I1227 08:08:54.925143 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9874\" (UniqueName: \"kubernetes.io/projected/14b9a667-b34c-4345-8c38-711f74fe9a59-kube-api-access-s9874\") pod \"dnsmasq-dns-5b75489c6f-8g7tc\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:54 crc kubenswrapper[4934]: I1227 08:08:54.925209 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b75489c6f-8g7tc\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:54 crc kubenswrapper[4934]: I1227 08:08:54.925226 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-ovsdbserver-nb\") pod \"dnsmasq-dns-5b75489c6f-8g7tc\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:54 crc kubenswrapper[4934]: I1227 08:08:54.925244 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-config\") pod \"dnsmasq-dns-5b75489c6f-8g7tc\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:55 crc kubenswrapper[4934]: I1227 08:08:55.027883 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b75489c6f-8g7tc\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:55 crc kubenswrapper[4934]: I1227 08:08:55.027925 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-ovsdbserver-nb\") pod \"dnsmasq-dns-5b75489c6f-8g7tc\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:55 crc kubenswrapper[4934]: I1227 08:08:55.027948 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-config\") pod \"dnsmasq-dns-5b75489c6f-8g7tc\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:55 crc kubenswrapper[4934]: I1227 08:08:55.028014 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-ovsdbserver-sb\") pod \"dnsmasq-dns-5b75489c6f-8g7tc\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:55 crc kubenswrapper[4934]: I1227 08:08:55.028093 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-dns-swift-storage-0\") pod \"dnsmasq-dns-5b75489c6f-8g7tc\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:55 crc kubenswrapper[4934]: I1227 08:08:55.028181 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-dns-svc\") pod \"dnsmasq-dns-5b75489c6f-8g7tc\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:55 crc kubenswrapper[4934]: I1227 08:08:55.028206 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9874\" (UniqueName: \"kubernetes.io/projected/14b9a667-b34c-4345-8c38-711f74fe9a59-kube-api-access-s9874\") pod \"dnsmasq-dns-5b75489c6f-8g7tc\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:55 crc kubenswrapper[4934]: I1227 08:08:55.028798 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b75489c6f-8g7tc\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:55 crc kubenswrapper[4934]: I1227 08:08:55.028828 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-config\") pod \"dnsmasq-dns-5b75489c6f-8g7tc\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:55 crc kubenswrapper[4934]: I1227 08:08:55.029154 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-dns-swift-storage-0\") pod \"dnsmasq-dns-5b75489c6f-8g7tc\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:55 crc kubenswrapper[4934]: I1227 08:08:55.029208 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-dns-svc\") pod \"dnsmasq-dns-5b75489c6f-8g7tc\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:55 crc kubenswrapper[4934]: I1227 08:08:55.029453 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-ovsdbserver-nb\") pod \"dnsmasq-dns-5b75489c6f-8g7tc\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:55 crc kubenswrapper[4934]: I1227 08:08:55.029733 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-ovsdbserver-sb\") pod \"dnsmasq-dns-5b75489c6f-8g7tc\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:55 crc kubenswrapper[4934]: I1227 08:08:55.051599 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9874\" (UniqueName: \"kubernetes.io/projected/14b9a667-b34c-4345-8c38-711f74fe9a59-kube-api-access-s9874\") pod \"dnsmasq-dns-5b75489c6f-8g7tc\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:55 crc kubenswrapper[4934]: I1227 08:08:55.185408 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:08:58 crc kubenswrapper[4934]: I1227 08:08:58.850209 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="df7d6b06-cb87-4635-8aca-1ab00a564dd7" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.131:5671: connect: connection refused" Dec 27 08:09:01 crc kubenswrapper[4934]: E1227 08:09:01.291775 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 27 08:09:01 crc kubenswrapper[4934]: E1227 08:09:01.292448 4934 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 27 08:09:01 crc kubenswrapper[4934]: E1227 08:09:01.292608 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tlvw8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-njzm7_openstack(45927aae-ef6c-475a-b48f-59cd34ff1278): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 08:09:01 crc kubenswrapper[4934]: E1227 08:09:01.293791 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-njzm7" podUID="45927aae-ef6c-475a-b48f-59cd34ff1278" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.429506 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.540882 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"76158d63-2099-4418-82a2-09af0432c1a3","Type":"ContainerDied","Data":"b8c7453eebadd180cef454914d2a20555fad716db7185abf26af4c945f82714d"} Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.540956 4934 scope.go:117] "RemoveContainer" containerID="d137c3d7f1fb61a5492ae4e87102e9b26b5caceeac796ed7e5db20c3505e981d" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.541211 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.542458 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-confd\") pod \"76158d63-2099-4418-82a2-09af0432c1a3\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.542641 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/76158d63-2099-4418-82a2-09af0432c1a3-pod-info\") pod \"76158d63-2099-4418-82a2-09af0432c1a3\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.542741 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/76158d63-2099-4418-82a2-09af0432c1a3-erlang-cookie-secret\") pod \"76158d63-2099-4418-82a2-09af0432c1a3\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.542870 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxr2j\" (UniqueName: \"kubernetes.io/projected/76158d63-2099-4418-82a2-09af0432c1a3-kube-api-access-lxr2j\") pod \"76158d63-2099-4418-82a2-09af0432c1a3\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.543042 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/76158d63-2099-4418-82a2-09af0432c1a3-server-conf\") pod \"76158d63-2099-4418-82a2-09af0432c1a3\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.543157 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-erlang-cookie\") pod \"76158d63-2099-4418-82a2-09af0432c1a3\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.543364 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76158d63-2099-4418-82a2-09af0432c1a3-config-data\") pod \"76158d63-2099-4418-82a2-09af0432c1a3\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.543471 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-tls\") pod \"76158d63-2099-4418-82a2-09af0432c1a3\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.544235 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e\") pod \"76158d63-2099-4418-82a2-09af0432c1a3\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.544408 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-plugins\") pod \"76158d63-2099-4418-82a2-09af0432c1a3\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.544781 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/76158d63-2099-4418-82a2-09af0432c1a3-plugins-conf\") pod \"76158d63-2099-4418-82a2-09af0432c1a3\" (UID: \"76158d63-2099-4418-82a2-09af0432c1a3\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.544850 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "76158d63-2099-4418-82a2-09af0432c1a3" (UID: "76158d63-2099-4418-82a2-09af0432c1a3"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.545978 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "76158d63-2099-4418-82a2-09af0432c1a3" (UID: "76158d63-2099-4418-82a2-09af0432c1a3"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.547495 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76158d63-2099-4418-82a2-09af0432c1a3-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "76158d63-2099-4418-82a2-09af0432c1a3" (UID: "76158d63-2099-4418-82a2-09af0432c1a3"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.547683 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.547766 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:01 crc kubenswrapper[4934]: E1227 08:09:01.548421 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-njzm7" podUID="45927aae-ef6c-475a-b48f-59cd34ff1278" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.550178 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/76158d63-2099-4418-82a2-09af0432c1a3-pod-info" (OuterVolumeSpecName: "pod-info") pod "76158d63-2099-4418-82a2-09af0432c1a3" (UID: "76158d63-2099-4418-82a2-09af0432c1a3"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.550212 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76158d63-2099-4418-82a2-09af0432c1a3-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "76158d63-2099-4418-82a2-09af0432c1a3" (UID: "76158d63-2099-4418-82a2-09af0432c1a3"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.567356 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "76158d63-2099-4418-82a2-09af0432c1a3" (UID: "76158d63-2099-4418-82a2-09af0432c1a3"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.573478 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76158d63-2099-4418-82a2-09af0432c1a3-kube-api-access-lxr2j" (OuterVolumeSpecName: "kube-api-access-lxr2j") pod "76158d63-2099-4418-82a2-09af0432c1a3" (UID: "76158d63-2099-4418-82a2-09af0432c1a3"). InnerVolumeSpecName "kube-api-access-lxr2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.635137 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e" (OuterVolumeSpecName: "persistence") pod "76158d63-2099-4418-82a2-09af0432c1a3" (UID: "76158d63-2099-4418-82a2-09af0432c1a3"). InnerVolumeSpecName "pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.637015 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76158d63-2099-4418-82a2-09af0432c1a3-config-data" (OuterVolumeSpecName: "config-data") pod "76158d63-2099-4418-82a2-09af0432c1a3" (UID: "76158d63-2099-4418-82a2-09af0432c1a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.663073 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e\") on node \"crc\" " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.663140 4934 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/76158d63-2099-4418-82a2-09af0432c1a3-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.663156 4934 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/76158d63-2099-4418-82a2-09af0432c1a3-pod-info\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.663170 4934 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/76158d63-2099-4418-82a2-09af0432c1a3-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.663184 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxr2j\" (UniqueName: \"kubernetes.io/projected/76158d63-2099-4418-82a2-09af0432c1a3-kube-api-access-lxr2j\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.663209 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76158d63-2099-4418-82a2-09af0432c1a3-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.663223 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.722320 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76158d63-2099-4418-82a2-09af0432c1a3-server-conf" (OuterVolumeSpecName: "server-conf") pod "76158d63-2099-4418-82a2-09af0432c1a3" (UID: "76158d63-2099-4418-82a2-09af0432c1a3"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.734260 4934 scope.go:117] "RemoveContainer" containerID="e6252db8825d6b9fe04ede9d278e6c3c2fa0bc3d3cf385da4ee6ce5043dc6b19" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.748643 4934 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.749030 4934 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e") on node "crc" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.753188 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.783243 4934 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/76158d63-2099-4418-82a2-09af0432c1a3-server-conf\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.783289 4934 reconciler_common.go:293] "Volume detached for volume \"pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.788897 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "76158d63-2099-4418-82a2-09af0432c1a3" (UID: "76158d63-2099-4418-82a2-09af0432c1a3"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.884646 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df7d6b06-cb87-4635-8aca-1ab00a564dd7-erlang-cookie-secret\") pod \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.884683 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df7d6b06-cb87-4635-8aca-1ab00a564dd7-server-conf\") pod \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.884720 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df7d6b06-cb87-4635-8aca-1ab00a564dd7-config-data\") pod \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.884800 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-erlang-cookie\") pod \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.884818 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-tls\") pod \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.884878 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-plugins\") pod \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.884991 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df7d6b06-cb87-4635-8aca-1ab00a564dd7-pod-info\") pod \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.885034 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df7d6b06-cb87-4635-8aca-1ab00a564dd7-plugins-conf\") pod \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.885128 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qr84v\" (UniqueName: \"kubernetes.io/projected/df7d6b06-cb87-4635-8aca-1ab00a564dd7-kube-api-access-qr84v\") pod \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.886487 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "df7d6b06-cb87-4635-8aca-1ab00a564dd7" (UID: "df7d6b06-cb87-4635-8aca-1ab00a564dd7"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.888149 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "df7d6b06-cb87-4635-8aca-1ab00a564dd7" (UID: "df7d6b06-cb87-4635-8aca-1ab00a564dd7"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.888253 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c561df02-e37e-4324-820c-98030eeb03e0\") pod \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.888440 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df7d6b06-cb87-4635-8aca-1ab00a564dd7-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "df7d6b06-cb87-4635-8aca-1ab00a564dd7" (UID: "df7d6b06-cb87-4635-8aca-1ab00a564dd7"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.888449 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-confd\") pod \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\" (UID: \"df7d6b06-cb87-4635-8aca-1ab00a564dd7\") " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.889504 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.889518 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/76158d63-2099-4418-82a2-09af0432c1a3-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.889527 4934 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df7d6b06-cb87-4635-8aca-1ab00a564dd7-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.889536 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.893752 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "df7d6b06-cb87-4635-8aca-1ab00a564dd7" (UID: "df7d6b06-cb87-4635-8aca-1ab00a564dd7"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.894934 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df7d6b06-cb87-4635-8aca-1ab00a564dd7-kube-api-access-qr84v" (OuterVolumeSpecName: "kube-api-access-qr84v") pod "df7d6b06-cb87-4635-8aca-1ab00a564dd7" (UID: "df7d6b06-cb87-4635-8aca-1ab00a564dd7"). InnerVolumeSpecName "kube-api-access-qr84v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.895706 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.910400 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.915194 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/df7d6b06-cb87-4635-8aca-1ab00a564dd7-pod-info" (OuterVolumeSpecName: "pod-info") pod "df7d6b06-cb87-4635-8aca-1ab00a564dd7" (UID: "df7d6b06-cb87-4635-8aca-1ab00a564dd7"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.921361 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df7d6b06-cb87-4635-8aca-1ab00a564dd7-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "df7d6b06-cb87-4635-8aca-1ab00a564dd7" (UID: "df7d6b06-cb87-4635-8aca-1ab00a564dd7"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.970296 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df7d6b06-cb87-4635-8aca-1ab00a564dd7-config-data" (OuterVolumeSpecName: "config-data") pod "df7d6b06-cb87-4635-8aca-1ab00a564dd7" (UID: "df7d6b06-cb87-4635-8aca-1ab00a564dd7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.981626 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 27 08:09:01 crc kubenswrapper[4934]: E1227 08:09:01.982180 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df7d6b06-cb87-4635-8aca-1ab00a564dd7" containerName="setup-container" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.982199 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="df7d6b06-cb87-4635-8aca-1ab00a564dd7" containerName="setup-container" Dec 27 08:09:01 crc kubenswrapper[4934]: E1227 08:09:01.982228 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76158d63-2099-4418-82a2-09af0432c1a3" containerName="setup-container" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.982234 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="76158d63-2099-4418-82a2-09af0432c1a3" containerName="setup-container" Dec 27 08:09:01 crc kubenswrapper[4934]: E1227 08:09:01.982247 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76158d63-2099-4418-82a2-09af0432c1a3" containerName="rabbitmq" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.982253 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="76158d63-2099-4418-82a2-09af0432c1a3" containerName="rabbitmq" Dec 27 08:09:01 crc kubenswrapper[4934]: E1227 08:09:01.982263 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df7d6b06-cb87-4635-8aca-1ab00a564dd7" containerName="rabbitmq" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.982269 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="df7d6b06-cb87-4635-8aca-1ab00a564dd7" containerName="rabbitmq" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.982499 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="df7d6b06-cb87-4635-8aca-1ab00a564dd7" containerName="rabbitmq" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.982540 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="76158d63-2099-4418-82a2-09af0432c1a3" containerName="rabbitmq" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.983786 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.985172 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c561df02-e37e-4324-820c-98030eeb03e0" (OuterVolumeSpecName: "persistence") pod "df7d6b06-cb87-4635-8aca-1ab00a564dd7" (UID: "df7d6b06-cb87-4635-8aca-1ab00a564dd7"). InnerVolumeSpecName "pvc-c561df02-e37e-4324-820c-98030eeb03e0". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.991938 4934 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df7d6b06-cb87-4635-8aca-1ab00a564dd7-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.991966 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df7d6b06-cb87-4635-8aca-1ab00a564dd7-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.991978 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.991986 4934 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df7d6b06-cb87-4635-8aca-1ab00a564dd7-pod-info\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.991995 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qr84v\" (UniqueName: \"kubernetes.io/projected/df7d6b06-cb87-4635-8aca-1ab00a564dd7-kube-api-access-qr84v\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.992018 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-c561df02-e37e-4324-820c-98030eeb03e0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c561df02-e37e-4324-820c-98030eeb03e0\") on node \"crc\" " Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.992284 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-kn2rk" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.992453 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.992602 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.992728 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.992841 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.992970 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.993228 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 27 08:09:01 crc kubenswrapper[4934]: I1227 08:09:01.997789 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.084686 4934 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.084917 4934 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-c561df02-e37e-4324-820c-98030eeb03e0" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c561df02-e37e-4324-820c-98030eeb03e0") on node "crc" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.085485 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df7d6b06-cb87-4635-8aca-1ab00a564dd7-server-conf" (OuterVolumeSpecName: "server-conf") pod "df7d6b06-cb87-4635-8aca-1ab00a564dd7" (UID: "df7d6b06-cb87-4635-8aca-1ab00a564dd7"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.095468 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.095950 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/85c58dd9-a367-4468-ad44-0f567a89c903-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.096106 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/85c58dd9-a367-4468-ad44-0f567a89c903-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.096128 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/85c58dd9-a367-4468-ad44-0f567a89c903-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.096150 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjwtg\" (UniqueName: \"kubernetes.io/projected/85c58dd9-a367-4468-ad44-0f567a89c903-kube-api-access-sjwtg\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.096189 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/85c58dd9-a367-4468-ad44-0f567a89c903-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.096233 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/85c58dd9-a367-4468-ad44-0f567a89c903-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.096264 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/85c58dd9-a367-4468-ad44-0f567a89c903-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.096441 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/85c58dd9-a367-4468-ad44-0f567a89c903-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.096477 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/85c58dd9-a367-4468-ad44-0f567a89c903-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.096556 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/85c58dd9-a367-4468-ad44-0f567a89c903-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.097194 4934 reconciler_common.go:293] "Volume detached for volume \"pvc-c561df02-e37e-4324-820c-98030eeb03e0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c561df02-e37e-4324-820c-98030eeb03e0\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.097228 4934 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df7d6b06-cb87-4635-8aca-1ab00a564dd7-server-conf\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.118590 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "df7d6b06-cb87-4635-8aca-1ab00a564dd7" (UID: "df7d6b06-cb87-4635-8aca-1ab00a564dd7"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:09:02 crc kubenswrapper[4934]: W1227 08:09:02.181292 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14b9a667_b34c_4345_8c38_711f74fe9a59.slice/crio-5e64e85517304f83f96f311296e9695d7b244296df4a419c990d831571f09361 WatchSource:0}: Error finding container 5e64e85517304f83f96f311296e9695d7b244296df4a419c990d831571f09361: Status 404 returned error can't find the container with id 5e64e85517304f83f96f311296e9695d7b244296df4a419c990d831571f09361 Dec 27 08:09:02 crc kubenswrapper[4934]: W1227 08:09:02.181731 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12e88dda_31ca_4b5b_93c3_7eced0cacda7.slice/crio-91aaee4c9203052c554a3b97535ad1b215e4e586b8bc7550ffaf876d032a7982 WatchSource:0}: Error finding container 91aaee4c9203052c554a3b97535ad1b215e4e586b8bc7550ffaf876d032a7982: Status 404 returned error can't find the container with id 91aaee4c9203052c554a3b97535ad1b215e4e586b8bc7550ffaf876d032a7982 Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.183657 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-8g7tc"] Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.197447 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.199238 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/85c58dd9-a367-4468-ad44-0f567a89c903-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.199291 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/85c58dd9-a367-4468-ad44-0f567a89c903-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.199328 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjwtg\" (UniqueName: \"kubernetes.io/projected/85c58dd9-a367-4468-ad44-0f567a89c903-kube-api-access-sjwtg\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.199365 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/85c58dd9-a367-4468-ad44-0f567a89c903-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.199405 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/85c58dd9-a367-4468-ad44-0f567a89c903-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.199446 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/85c58dd9-a367-4468-ad44-0f567a89c903-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.199555 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/85c58dd9-a367-4468-ad44-0f567a89c903-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.199711 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/85c58dd9-a367-4468-ad44-0f567a89c903-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.199798 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/85c58dd9-a367-4468-ad44-0f567a89c903-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.199876 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.199935 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/85c58dd9-a367-4468-ad44-0f567a89c903-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.200038 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df7d6b06-cb87-4635-8aca-1ab00a564dd7-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.200500 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/85c58dd9-a367-4468-ad44-0f567a89c903-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.200515 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/85c58dd9-a367-4468-ad44-0f567a89c903-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.201031 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/85c58dd9-a367-4468-ad44-0f567a89c903-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.201672 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/85c58dd9-a367-4468-ad44-0f567a89c903-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.203499 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/85c58dd9-a367-4468-ad44-0f567a89c903-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.203907 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/85c58dd9-a367-4468-ad44-0f567a89c903-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.204415 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/85c58dd9-a367-4468-ad44-0f567a89c903-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.204774 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/85c58dd9-a367-4468-ad44-0f567a89c903-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.207457 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.207500 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/65d60d7c646e54e3493361ab222a8aba7e9f991db4c177aebfe00d23097f5c94/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.213212 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/85c58dd9-a367-4468-ad44-0f567a89c903-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.216922 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjwtg\" (UniqueName: \"kubernetes.io/projected/85c58dd9-a367-4468-ad44-0f567a89c903-kube-api-access-sjwtg\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.281782 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cefc97b0-9636-43b6-9eb8-3128dde08d4e\") pod \"rabbitmq-cell1-server-0\" (UID: \"85c58dd9-a367-4468-ad44-0f567a89c903\") " pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.325951 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.574992 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"df7d6b06-cb87-4635-8aca-1ab00a564dd7","Type":"ContainerDied","Data":"4f79111b4aad5ba928b46f818c525852ae66ed2784c724af6d61b660165a3f00"} Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.575404 4934 scope.go:117] "RemoveContainer" containerID="7572ade40e346d8651cbf1ce8fabe437ad6fbf893725ab3cc22288f68c146c30" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.575668 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.579674 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12e88dda-31ca-4b5b-93c3-7eced0cacda7","Type":"ContainerStarted","Data":"91aaee4c9203052c554a3b97535ad1b215e4e586b8bc7550ffaf876d032a7982"} Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.584354 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" event={"ID":"14b9a667-b34c-4345-8c38-711f74fe9a59","Type":"ContainerStarted","Data":"5e64e85517304f83f96f311296e9695d7b244296df4a419c990d831571f09361"} Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.619059 4934 scope.go:117] "RemoveContainer" containerID="6386108486d5f92741eb22a890f1a957ebb07da8c2f70cbbed654caafe93c1c2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.656856 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-2"] Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.700633 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-2"] Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.722140 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-2"] Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.724221 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.756831 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-2"] Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.822473 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-config-data\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.822616 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-pod-info\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.822648 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.822688 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-server-conf\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.822707 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.822731 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l69dd\" (UniqueName: \"kubernetes.io/projected/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-kube-api-access-l69dd\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.822759 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.822777 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.822801 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c561df02-e37e-4324-820c-98030eeb03e0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c561df02-e37e-4324-820c-98030eeb03e0\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.822816 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.822846 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.916360 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.927919 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-server-conf\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.927959 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.927987 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l69dd\" (UniqueName: \"kubernetes.io/projected/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-kube-api-access-l69dd\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.928020 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.928039 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.928064 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c561df02-e37e-4324-820c-98030eeb03e0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c561df02-e37e-4324-820c-98030eeb03e0\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.928101 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.928138 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.928183 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-config-data\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.928276 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-pod-info\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.928300 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.930447 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-server-conf\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.930687 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.940091 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.940796 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-config-data\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.941857 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.949681 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-pod-info\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.952978 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.953434 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.953463 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c561df02-e37e-4324-820c-98030eeb03e0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c561df02-e37e-4324-820c-98030eeb03e0\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5d9ca5747749969804eb38b0c8f9036813747c72b7012090db08a3a954d2ec6d/globalmount\"" pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.953489 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.962805 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:02 crc kubenswrapper[4934]: I1227 08:09:02.974269 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l69dd\" (UniqueName: \"kubernetes.io/projected/5a3ecad3-f6df-4051-be94-faff5cbc4e1f-kube-api-access-l69dd\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:03 crc kubenswrapper[4934]: I1227 08:09:03.045419 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c561df02-e37e-4324-820c-98030eeb03e0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c561df02-e37e-4324-820c-98030eeb03e0\") pod \"rabbitmq-server-2\" (UID: \"5a3ecad3-f6df-4051-be94-faff5cbc4e1f\") " pod="openstack/rabbitmq-server-2" Dec 27 08:09:03 crc kubenswrapper[4934]: I1227 08:09:03.321642 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Dec 27 08:09:03 crc kubenswrapper[4934]: I1227 08:09:03.489767 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76158d63-2099-4418-82a2-09af0432c1a3" path="/var/lib/kubelet/pods/76158d63-2099-4418-82a2-09af0432c1a3/volumes" Dec 27 08:09:03 crc kubenswrapper[4934]: I1227 08:09:03.492884 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df7d6b06-cb87-4635-8aca-1ab00a564dd7" path="/var/lib/kubelet/pods/df7d6b06-cb87-4635-8aca-1ab00a564dd7/volumes" Dec 27 08:09:03 crc kubenswrapper[4934]: I1227 08:09:03.635630 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"85c58dd9-a367-4468-ad44-0f567a89c903","Type":"ContainerStarted","Data":"9a19f0f7ded31fd8cc4c73b393f41cb1260b65c24a1998c8ea5a91488ba1d61e"} Dec 27 08:09:03 crc kubenswrapper[4934]: I1227 08:09:03.647934 4934 generic.go:334] "Generic (PLEG): container finished" podID="14b9a667-b34c-4345-8c38-711f74fe9a59" containerID="8a817503de77493decb32cf22084df456afa80493ac0cc871509e1083c69ff0a" exitCode=0 Dec 27 08:09:03 crc kubenswrapper[4934]: I1227 08:09:03.647980 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" event={"ID":"14b9a667-b34c-4345-8c38-711f74fe9a59","Type":"ContainerDied","Data":"8a817503de77493decb32cf22084df456afa80493ac0cc871509e1083c69ff0a"} Dec 27 08:09:03 crc kubenswrapper[4934]: I1227 08:09:03.905387 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-2"] Dec 27 08:09:04 crc kubenswrapper[4934]: I1227 08:09:04.064886 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="76158d63-2099-4418-82a2-09af0432c1a3" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.132:5671: i/o timeout" Dec 27 08:09:04 crc kubenswrapper[4934]: I1227 08:09:04.684930 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"5a3ecad3-f6df-4051-be94-faff5cbc4e1f","Type":"ContainerStarted","Data":"042502e2a5ac1cd186b5c4153d4b2a9d992c7c8a7a7197974884c34a5a93e5a2"} Dec 27 08:09:04 crc kubenswrapper[4934]: I1227 08:09:04.690124 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" event={"ID":"14b9a667-b34c-4345-8c38-711f74fe9a59","Type":"ContainerStarted","Data":"fd77dab3895b7b3614b62a36ddcb69f7be7632d3e0336dd5e0f064b94d675676"} Dec 27 08:09:04 crc kubenswrapper[4934]: I1227 08:09:04.690288 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:09:04 crc kubenswrapper[4934]: I1227 08:09:04.722203 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" podStartSLOduration=10.722181769 podStartE2EDuration="10.722181769s" podCreationTimestamp="2025-12-27 08:08:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:09:04.714873948 +0000 UTC m=+1605.535314542" watchObservedRunningTime="2025-12-27 08:09:04.722181769 +0000 UTC m=+1605.542622383" Dec 27 08:09:05 crc kubenswrapper[4934]: I1227 08:09:05.705175 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"85c58dd9-a367-4468-ad44-0f567a89c903","Type":"ContainerStarted","Data":"795f5acf5725079964e7a78f9a2820082b4abd45fe0a050721438cd2d158857d"} Dec 27 08:09:06 crc kubenswrapper[4934]: I1227 08:09:06.721401 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"5a3ecad3-f6df-4051-be94-faff5cbc4e1f","Type":"ContainerStarted","Data":"2502034eeae116a80d839154687bc95479ad709015cfb7946d76d6ad6f13679a"} Dec 27 08:09:07 crc kubenswrapper[4934]: I1227 08:09:07.471163 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:09:07 crc kubenswrapper[4934]: E1227 08:09:07.472365 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:09:07 crc kubenswrapper[4934]: I1227 08:09:07.739231 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12e88dda-31ca-4b5b-93c3-7eced0cacda7","Type":"ContainerStarted","Data":"9a1e16b5a1faeb0c73061b916b68268113d66cfd5ed3b6d232dc6f1725c20d73"} Dec 27 08:09:07 crc kubenswrapper[4934]: I1227 08:09:07.739295 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12e88dda-31ca-4b5b-93c3-7eced0cacda7","Type":"ContainerStarted","Data":"9995c35f73f77d890191ce4b613fd43a168910498934a2d034fa50716244fb04"} Dec 27 08:09:08 crc kubenswrapper[4934]: I1227 08:09:08.764687 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12e88dda-31ca-4b5b-93c3-7eced0cacda7","Type":"ContainerStarted","Data":"ad0bb3214f6e29f8bf62f063a50e21a4b4a219cd599143f173a6e5e1b87f8532"} Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.187278 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.271623 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-mr67x"] Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.272063 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" podUID="1cede442-232c-45b8-a8d3-7d617531d53c" containerName="dnsmasq-dns" containerID="cri-o://012b79d6af6ea82699a9fade6fb224056c3ba52ab3a89131e4db688efd4fca9e" gracePeriod=10 Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.480034 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d75f767dc-pv2wg"] Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.492268 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.505318 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d75f767dc-pv2wg"] Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.596304 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e4329ea-695d-49e5-a530-932605451255-dns-svc\") pod \"dnsmasq-dns-5d75f767dc-pv2wg\" (UID: \"3e4329ea-695d-49e5-a530-932605451255\") " pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.597207 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e4329ea-695d-49e5-a530-932605451255-ovsdbserver-sb\") pod \"dnsmasq-dns-5d75f767dc-pv2wg\" (UID: \"3e4329ea-695d-49e5-a530-932605451255\") " pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.597691 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3e4329ea-695d-49e5-a530-932605451255-openstack-edpm-ipam\") pod \"dnsmasq-dns-5d75f767dc-pv2wg\" (UID: \"3e4329ea-695d-49e5-a530-932605451255\") " pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.597820 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e4329ea-695d-49e5-a530-932605451255-dns-swift-storage-0\") pod \"dnsmasq-dns-5d75f767dc-pv2wg\" (UID: \"3e4329ea-695d-49e5-a530-932605451255\") " pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.597960 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e4329ea-695d-49e5-a530-932605451255-ovsdbserver-nb\") pod \"dnsmasq-dns-5d75f767dc-pv2wg\" (UID: \"3e4329ea-695d-49e5-a530-932605451255\") " pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.598434 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj7n5\" (UniqueName: \"kubernetes.io/projected/3e4329ea-695d-49e5-a530-932605451255-kube-api-access-xj7n5\") pod \"dnsmasq-dns-5d75f767dc-pv2wg\" (UID: \"3e4329ea-695d-49e5-a530-932605451255\") " pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.598843 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e4329ea-695d-49e5-a530-932605451255-config\") pod \"dnsmasq-dns-5d75f767dc-pv2wg\" (UID: \"3e4329ea-695d-49e5-a530-932605451255\") " pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.700574 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e4329ea-695d-49e5-a530-932605451255-dns-svc\") pod \"dnsmasq-dns-5d75f767dc-pv2wg\" (UID: \"3e4329ea-695d-49e5-a530-932605451255\") " pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.700639 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e4329ea-695d-49e5-a530-932605451255-ovsdbserver-sb\") pod \"dnsmasq-dns-5d75f767dc-pv2wg\" (UID: \"3e4329ea-695d-49e5-a530-932605451255\") " pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.700671 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3e4329ea-695d-49e5-a530-932605451255-openstack-edpm-ipam\") pod \"dnsmasq-dns-5d75f767dc-pv2wg\" (UID: \"3e4329ea-695d-49e5-a530-932605451255\") " pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.700691 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e4329ea-695d-49e5-a530-932605451255-dns-swift-storage-0\") pod \"dnsmasq-dns-5d75f767dc-pv2wg\" (UID: \"3e4329ea-695d-49e5-a530-932605451255\") " pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.700732 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e4329ea-695d-49e5-a530-932605451255-ovsdbserver-nb\") pod \"dnsmasq-dns-5d75f767dc-pv2wg\" (UID: \"3e4329ea-695d-49e5-a530-932605451255\") " pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.700762 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj7n5\" (UniqueName: \"kubernetes.io/projected/3e4329ea-695d-49e5-a530-932605451255-kube-api-access-xj7n5\") pod \"dnsmasq-dns-5d75f767dc-pv2wg\" (UID: \"3e4329ea-695d-49e5-a530-932605451255\") " pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.700831 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e4329ea-695d-49e5-a530-932605451255-config\") pod \"dnsmasq-dns-5d75f767dc-pv2wg\" (UID: \"3e4329ea-695d-49e5-a530-932605451255\") " pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.701879 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e4329ea-695d-49e5-a530-932605451255-config\") pod \"dnsmasq-dns-5d75f767dc-pv2wg\" (UID: \"3e4329ea-695d-49e5-a530-932605451255\") " pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.702294 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e4329ea-695d-49e5-a530-932605451255-ovsdbserver-sb\") pod \"dnsmasq-dns-5d75f767dc-pv2wg\" (UID: \"3e4329ea-695d-49e5-a530-932605451255\") " pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.703662 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e4329ea-695d-49e5-a530-932605451255-dns-swift-storage-0\") pod \"dnsmasq-dns-5d75f767dc-pv2wg\" (UID: \"3e4329ea-695d-49e5-a530-932605451255\") " pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.704075 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e4329ea-695d-49e5-a530-932605451255-dns-svc\") pod \"dnsmasq-dns-5d75f767dc-pv2wg\" (UID: \"3e4329ea-695d-49e5-a530-932605451255\") " pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.705187 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3e4329ea-695d-49e5-a530-932605451255-openstack-edpm-ipam\") pod \"dnsmasq-dns-5d75f767dc-pv2wg\" (UID: \"3e4329ea-695d-49e5-a530-932605451255\") " pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.705495 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e4329ea-695d-49e5-a530-932605451255-ovsdbserver-nb\") pod \"dnsmasq-dns-5d75f767dc-pv2wg\" (UID: \"3e4329ea-695d-49e5-a530-932605451255\") " pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.720935 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj7n5\" (UniqueName: \"kubernetes.io/projected/3e4329ea-695d-49e5-a530-932605451255-kube-api-access-xj7n5\") pod \"dnsmasq-dns-5d75f767dc-pv2wg\" (UID: \"3e4329ea-695d-49e5-a530-932605451255\") " pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.827431 4934 generic.go:334] "Generic (PLEG): container finished" podID="1cede442-232c-45b8-a8d3-7d617531d53c" containerID="012b79d6af6ea82699a9fade6fb224056c3ba52ab3a89131e4db688efd4fca9e" exitCode=0 Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.827513 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" event={"ID":"1cede442-232c-45b8-a8d3-7d617531d53c","Type":"ContainerDied","Data":"012b79d6af6ea82699a9fade6fb224056c3ba52ab3a89131e4db688efd4fca9e"} Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.834399 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12e88dda-31ca-4b5b-93c3-7eced0cacda7","Type":"ContainerStarted","Data":"836924e3f43fe5210bf012ae497e36df6e8228dd024a1dc5ca8b5c2b559b00de"} Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.836002 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.863496 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.873226 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=16.950042052 podStartE2EDuration="24.873206266s" podCreationTimestamp="2025-12-27 08:08:46 +0000 UTC" firstStartedPulling="2025-12-27 08:09:02.184040508 +0000 UTC m=+1603.004481092" lastFinishedPulling="2025-12-27 08:09:10.107204672 +0000 UTC m=+1610.927645306" observedRunningTime="2025-12-27 08:09:10.862750127 +0000 UTC m=+1611.683190721" watchObservedRunningTime="2025-12-27 08:09:10.873206266 +0000 UTC m=+1611.693646860" Dec 27 08:09:10 crc kubenswrapper[4934]: I1227 08:09:10.908954 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.006434 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-ovsdbserver-sb\") pod \"1cede442-232c-45b8-a8d3-7d617531d53c\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.006660 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4s7ct\" (UniqueName: \"kubernetes.io/projected/1cede442-232c-45b8-a8d3-7d617531d53c-kube-api-access-4s7ct\") pod \"1cede442-232c-45b8-a8d3-7d617531d53c\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.006695 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-config\") pod \"1cede442-232c-45b8-a8d3-7d617531d53c\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.006732 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-dns-swift-storage-0\") pod \"1cede442-232c-45b8-a8d3-7d617531d53c\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.006842 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-dns-svc\") pod \"1cede442-232c-45b8-a8d3-7d617531d53c\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.006909 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-ovsdbserver-nb\") pod \"1cede442-232c-45b8-a8d3-7d617531d53c\" (UID: \"1cede442-232c-45b8-a8d3-7d617531d53c\") " Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.018300 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cede442-232c-45b8-a8d3-7d617531d53c-kube-api-access-4s7ct" (OuterVolumeSpecName: "kube-api-access-4s7ct") pod "1cede442-232c-45b8-a8d3-7d617531d53c" (UID: "1cede442-232c-45b8-a8d3-7d617531d53c"). InnerVolumeSpecName "kube-api-access-4s7ct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.111545 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4s7ct\" (UniqueName: \"kubernetes.io/projected/1cede442-232c-45b8-a8d3-7d617531d53c-kube-api-access-4s7ct\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.124265 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-config" (OuterVolumeSpecName: "config") pod "1cede442-232c-45b8-a8d3-7d617531d53c" (UID: "1cede442-232c-45b8-a8d3-7d617531d53c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.182818 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1cede442-232c-45b8-a8d3-7d617531d53c" (UID: "1cede442-232c-45b8-a8d3-7d617531d53c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.188056 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1cede442-232c-45b8-a8d3-7d617531d53c" (UID: "1cede442-232c-45b8-a8d3-7d617531d53c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.204973 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1cede442-232c-45b8-a8d3-7d617531d53c" (UID: "1cede442-232c-45b8-a8d3-7d617531d53c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.206535 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1cede442-232c-45b8-a8d3-7d617531d53c" (UID: "1cede442-232c-45b8-a8d3-7d617531d53c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.213823 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.213852 4934 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.213871 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.213881 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.213890 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1cede442-232c-45b8-a8d3-7d617531d53c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.564433 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d75f767dc-pv2wg"] Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.850370 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" event={"ID":"1cede442-232c-45b8-a8d3-7d617531d53c","Type":"ContainerDied","Data":"2ee369df8613cf716d77563c55db97d2550da30e720365dbdbd0e2852227d276"} Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.851094 4934 scope.go:117] "RemoveContainer" containerID="012b79d6af6ea82699a9fade6fb224056c3ba52ab3a89131e4db688efd4fca9e" Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.851344 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-mr67x" Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.855386 4934 generic.go:334] "Generic (PLEG): container finished" podID="3e4329ea-695d-49e5-a530-932605451255" containerID="aa43e8f87ba4c0d4c44ab1c7e5a2854d41067df51c829967f6ee132dd53fe822" exitCode=0 Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.856693 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" event={"ID":"3e4329ea-695d-49e5-a530-932605451255","Type":"ContainerDied","Data":"aa43e8f87ba4c0d4c44ab1c7e5a2854d41067df51c829967f6ee132dd53fe822"} Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.856795 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" event={"ID":"3e4329ea-695d-49e5-a530-932605451255","Type":"ContainerStarted","Data":"b3b939701b9edc36cbfe2ac2283fa5bdc97e97fe73c3dd81f3fb4e2a78cd5d32"} Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.906341 4934 scope.go:117] "RemoveContainer" containerID="b5ea68957cea70f159ffd8d767539802def518bcb94ba522afc2a1da848e2497" Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.922066 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-mr67x"] Dec 27 08:09:11 crc kubenswrapper[4934]: I1227 08:09:11.932974 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-mr67x"] Dec 27 08:09:12 crc kubenswrapper[4934]: I1227 08:09:12.872783 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" event={"ID":"3e4329ea-695d-49e5-a530-932605451255","Type":"ContainerStarted","Data":"4c97af197d4403da1195e1e752206f5e6b239eaf0948d15298960d39f5c15ed2"} Dec 27 08:09:12 crc kubenswrapper[4934]: I1227 08:09:12.902199 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" podStartSLOduration=2.902177634 podStartE2EDuration="2.902177634s" podCreationTimestamp="2025-12-27 08:09:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:09:12.893900249 +0000 UTC m=+1613.714340873" watchObservedRunningTime="2025-12-27 08:09:12.902177634 +0000 UTC m=+1613.722618268" Dec 27 08:09:13 crc kubenswrapper[4934]: I1227 08:09:13.482223 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cede442-232c-45b8-a8d3-7d617531d53c" path="/var/lib/kubelet/pods/1cede442-232c-45b8-a8d3-7d617531d53c/volumes" Dec 27 08:09:13 crc kubenswrapper[4934]: I1227 08:09:13.891451 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:15 crc kubenswrapper[4934]: I1227 08:09:15.935256 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-njzm7" event={"ID":"45927aae-ef6c-475a-b48f-59cd34ff1278","Type":"ContainerStarted","Data":"3a8c5484b16cfc1e465ec7b486a2604f58342498ccc50b8493e5c4578f77417d"} Dec 27 08:09:15 crc kubenswrapper[4934]: I1227 08:09:15.970632 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-njzm7" podStartSLOduration=2.536844945 podStartE2EDuration="38.970600422s" podCreationTimestamp="2025-12-27 08:08:37 +0000 UTC" firstStartedPulling="2025-12-27 08:08:38.404350767 +0000 UTC m=+1579.224791351" lastFinishedPulling="2025-12-27 08:09:14.838106194 +0000 UTC m=+1615.658546828" observedRunningTime="2025-12-27 08:09:15.951511298 +0000 UTC m=+1616.771951892" watchObservedRunningTime="2025-12-27 08:09:15.970600422 +0000 UTC m=+1616.791041056" Dec 27 08:09:17 crc kubenswrapper[4934]: I1227 08:09:17.966008 4934 generic.go:334] "Generic (PLEG): container finished" podID="45927aae-ef6c-475a-b48f-59cd34ff1278" containerID="3a8c5484b16cfc1e465ec7b486a2604f58342498ccc50b8493e5c4578f77417d" exitCode=0 Dec 27 08:09:17 crc kubenswrapper[4934]: I1227 08:09:17.966145 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-njzm7" event={"ID":"45927aae-ef6c-475a-b48f-59cd34ff1278","Type":"ContainerDied","Data":"3a8c5484b16cfc1e465ec7b486a2604f58342498ccc50b8493e5c4578f77417d"} Dec 27 08:09:18 crc kubenswrapper[4934]: I1227 08:09:18.469056 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:09:18 crc kubenswrapper[4934]: E1227 08:09:18.469954 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:09:19 crc kubenswrapper[4934]: I1227 08:09:19.505100 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-njzm7" Dec 27 08:09:19 crc kubenswrapper[4934]: I1227 08:09:19.670843 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45927aae-ef6c-475a-b48f-59cd34ff1278-config-data\") pod \"45927aae-ef6c-475a-b48f-59cd34ff1278\" (UID: \"45927aae-ef6c-475a-b48f-59cd34ff1278\") " Dec 27 08:09:19 crc kubenswrapper[4934]: I1227 08:09:19.671239 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlvw8\" (UniqueName: \"kubernetes.io/projected/45927aae-ef6c-475a-b48f-59cd34ff1278-kube-api-access-tlvw8\") pod \"45927aae-ef6c-475a-b48f-59cd34ff1278\" (UID: \"45927aae-ef6c-475a-b48f-59cd34ff1278\") " Dec 27 08:09:19 crc kubenswrapper[4934]: I1227 08:09:19.671577 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45927aae-ef6c-475a-b48f-59cd34ff1278-combined-ca-bundle\") pod \"45927aae-ef6c-475a-b48f-59cd34ff1278\" (UID: \"45927aae-ef6c-475a-b48f-59cd34ff1278\") " Dec 27 08:09:19 crc kubenswrapper[4934]: I1227 08:09:19.692431 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45927aae-ef6c-475a-b48f-59cd34ff1278-kube-api-access-tlvw8" (OuterVolumeSpecName: "kube-api-access-tlvw8") pod "45927aae-ef6c-475a-b48f-59cd34ff1278" (UID: "45927aae-ef6c-475a-b48f-59cd34ff1278"). InnerVolumeSpecName "kube-api-access-tlvw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:09:19 crc kubenswrapper[4934]: I1227 08:09:19.724331 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45927aae-ef6c-475a-b48f-59cd34ff1278-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "45927aae-ef6c-475a-b48f-59cd34ff1278" (UID: "45927aae-ef6c-475a-b48f-59cd34ff1278"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:09:19 crc kubenswrapper[4934]: I1227 08:09:19.774921 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlvw8\" (UniqueName: \"kubernetes.io/projected/45927aae-ef6c-475a-b48f-59cd34ff1278-kube-api-access-tlvw8\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:19 crc kubenswrapper[4934]: I1227 08:09:19.774962 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45927aae-ef6c-475a-b48f-59cd34ff1278-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:19 crc kubenswrapper[4934]: I1227 08:09:19.791765 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45927aae-ef6c-475a-b48f-59cd34ff1278-config-data" (OuterVolumeSpecName: "config-data") pod "45927aae-ef6c-475a-b48f-59cd34ff1278" (UID: "45927aae-ef6c-475a-b48f-59cd34ff1278"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:09:19 crc kubenswrapper[4934]: I1227 08:09:19.877759 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45927aae-ef6c-475a-b48f-59cd34ff1278-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:19 crc kubenswrapper[4934]: I1227 08:09:19.999142 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-njzm7" event={"ID":"45927aae-ef6c-475a-b48f-59cd34ff1278","Type":"ContainerDied","Data":"367a569da7ac5ff50eaf33b34df280128fff893009c17780dde69a5c99dca229"} Dec 27 08:09:19 crc kubenswrapper[4934]: I1227 08:09:19.999196 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="367a569da7ac5ff50eaf33b34df280128fff893009c17780dde69a5c99dca229" Dec 27 08:09:19 crc kubenswrapper[4934]: I1227 08:09:19.999222 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-njzm7" Dec 27 08:09:20 crc kubenswrapper[4934]: I1227 08:09:20.865323 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d75f767dc-pv2wg" Dec 27 08:09:20 crc kubenswrapper[4934]: I1227 08:09:20.957560 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-8g7tc"] Dec 27 08:09:20 crc kubenswrapper[4934]: I1227 08:09:20.958352 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" podUID="14b9a667-b34c-4345-8c38-711f74fe9a59" containerName="dnsmasq-dns" containerID="cri-o://fd77dab3895b7b3614b62a36ddcb69f7be7632d3e0336dd5e0f064b94d675676" gracePeriod=10 Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.396446 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-55c9c9fc75-s5tvt"] Dec 27 08:09:21 crc kubenswrapper[4934]: E1227 08:09:21.397057 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cede442-232c-45b8-a8d3-7d617531d53c" containerName="init" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.397075 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cede442-232c-45b8-a8d3-7d617531d53c" containerName="init" Dec 27 08:09:21 crc kubenswrapper[4934]: E1227 08:09:21.397100 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cede442-232c-45b8-a8d3-7d617531d53c" containerName="dnsmasq-dns" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.397107 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cede442-232c-45b8-a8d3-7d617531d53c" containerName="dnsmasq-dns" Dec 27 08:09:21 crc kubenswrapper[4934]: E1227 08:09:21.397136 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45927aae-ef6c-475a-b48f-59cd34ff1278" containerName="heat-db-sync" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.397144 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="45927aae-ef6c-475a-b48f-59cd34ff1278" containerName="heat-db-sync" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.397387 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cede442-232c-45b8-a8d3-7d617531d53c" containerName="dnsmasq-dns" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.397418 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="45927aae-ef6c-475a-b48f-59cd34ff1278" containerName="heat-db-sync" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.398203 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-55c9c9fc75-s5tvt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.424545 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-55c9c9fc75-s5tvt"] Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.443734 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb1fd5db-ac54-4e34-9cb3-4caae5f67da8-config-data-custom\") pod \"heat-engine-55c9c9fc75-s5tvt\" (UID: \"bb1fd5db-ac54-4e34-9cb3-4caae5f67da8\") " pod="openstack/heat-engine-55c9c9fc75-s5tvt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.443987 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb1fd5db-ac54-4e34-9cb3-4caae5f67da8-config-data\") pod \"heat-engine-55c9c9fc75-s5tvt\" (UID: \"bb1fd5db-ac54-4e34-9cb3-4caae5f67da8\") " pod="openstack/heat-engine-55c9c9fc75-s5tvt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.444122 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb1fd5db-ac54-4e34-9cb3-4caae5f67da8-combined-ca-bundle\") pod \"heat-engine-55c9c9fc75-s5tvt\" (UID: \"bb1fd5db-ac54-4e34-9cb3-4caae5f67da8\") " pod="openstack/heat-engine-55c9c9fc75-s5tvt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.444323 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll77j\" (UniqueName: \"kubernetes.io/projected/bb1fd5db-ac54-4e34-9cb3-4caae5f67da8-kube-api-access-ll77j\") pod \"heat-engine-55c9c9fc75-s5tvt\" (UID: \"bb1fd5db-ac54-4e34-9cb3-4caae5f67da8\") " pod="openstack/heat-engine-55c9c9fc75-s5tvt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.517122 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-8cb7b99d5-5xbmt"] Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.518710 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.533435 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-64f664dd6d-jgmlc"] Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.535034 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.548203 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb1fd5db-ac54-4e34-9cb3-4caae5f67da8-config-data-custom\") pod \"heat-engine-55c9c9fc75-s5tvt\" (UID: \"bb1fd5db-ac54-4e34-9cb3-4caae5f67da8\") " pod="openstack/heat-engine-55c9c9fc75-s5tvt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.548537 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb1fd5db-ac54-4e34-9cb3-4caae5f67da8-config-data\") pod \"heat-engine-55c9c9fc75-s5tvt\" (UID: \"bb1fd5db-ac54-4e34-9cb3-4caae5f67da8\") " pod="openstack/heat-engine-55c9c9fc75-s5tvt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.548621 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb1fd5db-ac54-4e34-9cb3-4caae5f67da8-combined-ca-bundle\") pod \"heat-engine-55c9c9fc75-s5tvt\" (UID: \"bb1fd5db-ac54-4e34-9cb3-4caae5f67da8\") " pod="openstack/heat-engine-55c9c9fc75-s5tvt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.548855 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll77j\" (UniqueName: \"kubernetes.io/projected/bb1fd5db-ac54-4e34-9cb3-4caae5f67da8-kube-api-access-ll77j\") pod \"heat-engine-55c9c9fc75-s5tvt\" (UID: \"bb1fd5db-ac54-4e34-9cb3-4caae5f67da8\") " pod="openstack/heat-engine-55c9c9fc75-s5tvt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.555498 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-8cb7b99d5-5xbmt"] Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.556365 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb1fd5db-ac54-4e34-9cb3-4caae5f67da8-config-data\") pod \"heat-engine-55c9c9fc75-s5tvt\" (UID: \"bb1fd5db-ac54-4e34-9cb3-4caae5f67da8\") " pod="openstack/heat-engine-55c9c9fc75-s5tvt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.557806 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb1fd5db-ac54-4e34-9cb3-4caae5f67da8-combined-ca-bundle\") pod \"heat-engine-55c9c9fc75-s5tvt\" (UID: \"bb1fd5db-ac54-4e34-9cb3-4caae5f67da8\") " pod="openstack/heat-engine-55c9c9fc75-s5tvt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.567404 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb1fd5db-ac54-4e34-9cb3-4caae5f67da8-config-data-custom\") pod \"heat-engine-55c9c9fc75-s5tvt\" (UID: \"bb1fd5db-ac54-4e34-9cb3-4caae5f67da8\") " pod="openstack/heat-engine-55c9c9fc75-s5tvt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.568043 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll77j\" (UniqueName: \"kubernetes.io/projected/bb1fd5db-ac54-4e34-9cb3-4caae5f67da8-kube-api-access-ll77j\") pod \"heat-engine-55c9c9fc75-s5tvt\" (UID: \"bb1fd5db-ac54-4e34-9cb3-4caae5f67da8\") " pod="openstack/heat-engine-55c9c9fc75-s5tvt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.570175 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-64f664dd6d-jgmlc"] Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.650707 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dklk\" (UniqueName: \"kubernetes.io/projected/2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43-kube-api-access-2dklk\") pod \"heat-api-64f664dd6d-jgmlc\" (UID: \"2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43\") " pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.650863 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43-config-data-custom\") pod \"heat-api-64f664dd6d-jgmlc\" (UID: \"2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43\") " pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.650987 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f10e6bef-1760-4a64-a702-675932d1770c-public-tls-certs\") pod \"heat-cfnapi-8cb7b99d5-5xbmt\" (UID: \"f10e6bef-1760-4a64-a702-675932d1770c\") " pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.652770 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f10e6bef-1760-4a64-a702-675932d1770c-config-data-custom\") pod \"heat-cfnapi-8cb7b99d5-5xbmt\" (UID: \"f10e6bef-1760-4a64-a702-675932d1770c\") " pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.652911 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43-public-tls-certs\") pod \"heat-api-64f664dd6d-jgmlc\" (UID: \"2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43\") " pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.652997 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f10e6bef-1760-4a64-a702-675932d1770c-internal-tls-certs\") pod \"heat-cfnapi-8cb7b99d5-5xbmt\" (UID: \"f10e6bef-1760-4a64-a702-675932d1770c\") " pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.653074 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43-internal-tls-certs\") pod \"heat-api-64f664dd6d-jgmlc\" (UID: \"2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43\") " pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.653253 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43-config-data\") pod \"heat-api-64f664dd6d-jgmlc\" (UID: \"2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43\") " pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.653327 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f10e6bef-1760-4a64-a702-675932d1770c-combined-ca-bundle\") pod \"heat-cfnapi-8cb7b99d5-5xbmt\" (UID: \"f10e6bef-1760-4a64-a702-675932d1770c\") " pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.653391 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10e6bef-1760-4a64-a702-675932d1770c-config-data\") pod \"heat-cfnapi-8cb7b99d5-5xbmt\" (UID: \"f10e6bef-1760-4a64-a702-675932d1770c\") " pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.653456 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mz65\" (UniqueName: \"kubernetes.io/projected/f10e6bef-1760-4a64-a702-675932d1770c-kube-api-access-6mz65\") pod \"heat-cfnapi-8cb7b99d5-5xbmt\" (UID: \"f10e6bef-1760-4a64-a702-675932d1770c\") " pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.653580 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43-combined-ca-bundle\") pod \"heat-api-64f664dd6d-jgmlc\" (UID: \"2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43\") " pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.718199 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-55c9c9fc75-s5tvt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.755156 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43-combined-ca-bundle\") pod \"heat-api-64f664dd6d-jgmlc\" (UID: \"2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43\") " pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.755374 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dklk\" (UniqueName: \"kubernetes.io/projected/2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43-kube-api-access-2dklk\") pod \"heat-api-64f664dd6d-jgmlc\" (UID: \"2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43\") " pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.756392 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43-config-data-custom\") pod \"heat-api-64f664dd6d-jgmlc\" (UID: \"2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43\") " pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.756874 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f10e6bef-1760-4a64-a702-675932d1770c-public-tls-certs\") pod \"heat-cfnapi-8cb7b99d5-5xbmt\" (UID: \"f10e6bef-1760-4a64-a702-675932d1770c\") " pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.757035 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f10e6bef-1760-4a64-a702-675932d1770c-config-data-custom\") pod \"heat-cfnapi-8cb7b99d5-5xbmt\" (UID: \"f10e6bef-1760-4a64-a702-675932d1770c\") " pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.757389 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43-public-tls-certs\") pod \"heat-api-64f664dd6d-jgmlc\" (UID: \"2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43\") " pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.757493 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f10e6bef-1760-4a64-a702-675932d1770c-internal-tls-certs\") pod \"heat-cfnapi-8cb7b99d5-5xbmt\" (UID: \"f10e6bef-1760-4a64-a702-675932d1770c\") " pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.757568 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43-internal-tls-certs\") pod \"heat-api-64f664dd6d-jgmlc\" (UID: \"2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43\") " pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.757656 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43-config-data\") pod \"heat-api-64f664dd6d-jgmlc\" (UID: \"2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43\") " pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.758420 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f10e6bef-1760-4a64-a702-675932d1770c-combined-ca-bundle\") pod \"heat-cfnapi-8cb7b99d5-5xbmt\" (UID: \"f10e6bef-1760-4a64-a702-675932d1770c\") " pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.758506 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10e6bef-1760-4a64-a702-675932d1770c-config-data\") pod \"heat-cfnapi-8cb7b99d5-5xbmt\" (UID: \"f10e6bef-1760-4a64-a702-675932d1770c\") " pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.758587 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mz65\" (UniqueName: \"kubernetes.io/projected/f10e6bef-1760-4a64-a702-675932d1770c-kube-api-access-6mz65\") pod \"heat-cfnapi-8cb7b99d5-5xbmt\" (UID: \"f10e6bef-1760-4a64-a702-675932d1770c\") " pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.761584 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43-public-tls-certs\") pod \"heat-api-64f664dd6d-jgmlc\" (UID: \"2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43\") " pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.761918 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43-combined-ca-bundle\") pod \"heat-api-64f664dd6d-jgmlc\" (UID: \"2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43\") " pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.763620 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f10e6bef-1760-4a64-a702-675932d1770c-public-tls-certs\") pod \"heat-cfnapi-8cb7b99d5-5xbmt\" (UID: \"f10e6bef-1760-4a64-a702-675932d1770c\") " pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.764389 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43-config-data-custom\") pod \"heat-api-64f664dd6d-jgmlc\" (UID: \"2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43\") " pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.765572 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43-internal-tls-certs\") pod \"heat-api-64f664dd6d-jgmlc\" (UID: \"2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43\") " pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.766872 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f10e6bef-1760-4a64-a702-675932d1770c-config-data-custom\") pod \"heat-cfnapi-8cb7b99d5-5xbmt\" (UID: \"f10e6bef-1760-4a64-a702-675932d1770c\") " pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.767479 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43-config-data\") pod \"heat-api-64f664dd6d-jgmlc\" (UID: \"2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43\") " pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.768270 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f10e6bef-1760-4a64-a702-675932d1770c-internal-tls-certs\") pod \"heat-cfnapi-8cb7b99d5-5xbmt\" (UID: \"f10e6bef-1760-4a64-a702-675932d1770c\") " pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.774408 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f10e6bef-1760-4a64-a702-675932d1770c-config-data\") pod \"heat-cfnapi-8cb7b99d5-5xbmt\" (UID: \"f10e6bef-1760-4a64-a702-675932d1770c\") " pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.775426 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mz65\" (UniqueName: \"kubernetes.io/projected/f10e6bef-1760-4a64-a702-675932d1770c-kube-api-access-6mz65\") pod \"heat-cfnapi-8cb7b99d5-5xbmt\" (UID: \"f10e6bef-1760-4a64-a702-675932d1770c\") " pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.776698 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dklk\" (UniqueName: \"kubernetes.io/projected/2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43-kube-api-access-2dklk\") pod \"heat-api-64f664dd6d-jgmlc\" (UID: \"2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43\") " pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.783172 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f10e6bef-1760-4a64-a702-675932d1770c-combined-ca-bundle\") pod \"heat-cfnapi-8cb7b99d5-5xbmt\" (UID: \"f10e6bef-1760-4a64-a702-675932d1770c\") " pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.840098 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.850861 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:21 crc kubenswrapper[4934]: I1227 08:09:21.976245 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.046200 4934 generic.go:334] "Generic (PLEG): container finished" podID="14b9a667-b34c-4345-8c38-711f74fe9a59" containerID="fd77dab3895b7b3614b62a36ddcb69f7be7632d3e0336dd5e0f064b94d675676" exitCode=0 Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.046241 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" event={"ID":"14b9a667-b34c-4345-8c38-711f74fe9a59","Type":"ContainerDied","Data":"fd77dab3895b7b3614b62a36ddcb69f7be7632d3e0336dd5e0f064b94d675676"} Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.046266 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" event={"ID":"14b9a667-b34c-4345-8c38-711f74fe9a59","Type":"ContainerDied","Data":"5e64e85517304f83f96f311296e9695d7b244296df4a419c990d831571f09361"} Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.046284 4934 scope.go:117] "RemoveContainer" containerID="fd77dab3895b7b3614b62a36ddcb69f7be7632d3e0336dd5e0f064b94d675676" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.046415 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-8g7tc" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.076237 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-openstack-edpm-ipam\") pod \"14b9a667-b34c-4345-8c38-711f74fe9a59\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.076310 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-config\") pod \"14b9a667-b34c-4345-8c38-711f74fe9a59\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.076491 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-dns-svc\") pod \"14b9a667-b34c-4345-8c38-711f74fe9a59\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.076523 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-dns-swift-storage-0\") pod \"14b9a667-b34c-4345-8c38-711f74fe9a59\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.076544 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-ovsdbserver-sb\") pod \"14b9a667-b34c-4345-8c38-711f74fe9a59\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.076700 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-ovsdbserver-nb\") pod \"14b9a667-b34c-4345-8c38-711f74fe9a59\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.076738 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9874\" (UniqueName: \"kubernetes.io/projected/14b9a667-b34c-4345-8c38-711f74fe9a59-kube-api-access-s9874\") pod \"14b9a667-b34c-4345-8c38-711f74fe9a59\" (UID: \"14b9a667-b34c-4345-8c38-711f74fe9a59\") " Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.083119 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14b9a667-b34c-4345-8c38-711f74fe9a59-kube-api-access-s9874" (OuterVolumeSpecName: "kube-api-access-s9874") pod "14b9a667-b34c-4345-8c38-711f74fe9a59" (UID: "14b9a667-b34c-4345-8c38-711f74fe9a59"). InnerVolumeSpecName "kube-api-access-s9874". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.086435 4934 scope.go:117] "RemoveContainer" containerID="8a817503de77493decb32cf22084df456afa80493ac0cc871509e1083c69ff0a" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.168053 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "14b9a667-b34c-4345-8c38-711f74fe9a59" (UID: "14b9a667-b34c-4345-8c38-711f74fe9a59"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.168795 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "14b9a667-b34c-4345-8c38-711f74fe9a59" (UID: "14b9a667-b34c-4345-8c38-711f74fe9a59"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.180286 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9874\" (UniqueName: \"kubernetes.io/projected/14b9a667-b34c-4345-8c38-711f74fe9a59-kube-api-access-s9874\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.180317 4934 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.180326 4934 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.183032 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "14b9a667-b34c-4345-8c38-711f74fe9a59" (UID: "14b9a667-b34c-4345-8c38-711f74fe9a59"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.204995 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "14b9a667-b34c-4345-8c38-711f74fe9a59" (UID: "14b9a667-b34c-4345-8c38-711f74fe9a59"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.237032 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-config" (OuterVolumeSpecName: "config") pod "14b9a667-b34c-4345-8c38-711f74fe9a59" (UID: "14b9a667-b34c-4345-8c38-711f74fe9a59"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.267280 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-55c9c9fc75-s5tvt"] Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.272861 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "14b9a667-b34c-4345-8c38-711f74fe9a59" (UID: "14b9a667-b34c-4345-8c38-711f74fe9a59"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.281926 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.281960 4934 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-config\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.281971 4934 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.281982 4934 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14b9a667-b34c-4345-8c38-711f74fe9a59-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.381631 4934 scope.go:117] "RemoveContainer" containerID="fd77dab3895b7b3614b62a36ddcb69f7be7632d3e0336dd5e0f064b94d675676" Dec 27 08:09:22 crc kubenswrapper[4934]: E1227 08:09:22.382406 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd77dab3895b7b3614b62a36ddcb69f7be7632d3e0336dd5e0f064b94d675676\": container with ID starting with fd77dab3895b7b3614b62a36ddcb69f7be7632d3e0336dd5e0f064b94d675676 not found: ID does not exist" containerID="fd77dab3895b7b3614b62a36ddcb69f7be7632d3e0336dd5e0f064b94d675676" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.382440 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd77dab3895b7b3614b62a36ddcb69f7be7632d3e0336dd5e0f064b94d675676"} err="failed to get container status \"fd77dab3895b7b3614b62a36ddcb69f7be7632d3e0336dd5e0f064b94d675676\": rpc error: code = NotFound desc = could not find container \"fd77dab3895b7b3614b62a36ddcb69f7be7632d3e0336dd5e0f064b94d675676\": container with ID starting with fd77dab3895b7b3614b62a36ddcb69f7be7632d3e0336dd5e0f064b94d675676 not found: ID does not exist" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.384364 4934 scope.go:117] "RemoveContainer" containerID="8a817503de77493decb32cf22084df456afa80493ac0cc871509e1083c69ff0a" Dec 27 08:09:22 crc kubenswrapper[4934]: E1227 08:09:22.384769 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a817503de77493decb32cf22084df456afa80493ac0cc871509e1083c69ff0a\": container with ID starting with 8a817503de77493decb32cf22084df456afa80493ac0cc871509e1083c69ff0a not found: ID does not exist" containerID="8a817503de77493decb32cf22084df456afa80493ac0cc871509e1083c69ff0a" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.384822 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a817503de77493decb32cf22084df456afa80493ac0cc871509e1083c69ff0a"} err="failed to get container status \"8a817503de77493decb32cf22084df456afa80493ac0cc871509e1083c69ff0a\": rpc error: code = NotFound desc = could not find container \"8a817503de77493decb32cf22084df456afa80493ac0cc871509e1083c69ff0a\": container with ID starting with 8a817503de77493decb32cf22084df456afa80493ac0cc871509e1083c69ff0a not found: ID does not exist" Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.421285 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-8g7tc"] Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.432905 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-8g7tc"] Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.445034 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-64f664dd6d-jgmlc"] Dec 27 08:09:22 crc kubenswrapper[4934]: W1227 08:09:22.448727 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ccf769c_f8ee_4c19_9ee7_1cd7d839ab43.slice/crio-d456be41188880d574bc5ece6d790c25db17cd299aed278d76c49136b64e2e03 WatchSource:0}: Error finding container d456be41188880d574bc5ece6d790c25db17cd299aed278d76c49136b64e2e03: Status 404 returned error can't find the container with id d456be41188880d574bc5ece6d790c25db17cd299aed278d76c49136b64e2e03 Dec 27 08:09:22 crc kubenswrapper[4934]: I1227 08:09:22.619915 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-8cb7b99d5-5xbmt"] Dec 27 08:09:22 crc kubenswrapper[4934]: W1227 08:09:22.622886 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf10e6bef_1760_4a64_a702_675932d1770c.slice/crio-bd8de5cb0eff83e5c0bc56a140221cfc955a3826c76ced4b78a8f182115ce5f4 WatchSource:0}: Error finding container bd8de5cb0eff83e5c0bc56a140221cfc955a3826c76ced4b78a8f182115ce5f4: Status 404 returned error can't find the container with id bd8de5cb0eff83e5c0bc56a140221cfc955a3826c76ced4b78a8f182115ce5f4 Dec 27 08:09:23 crc kubenswrapper[4934]: I1227 08:09:23.067505 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" event={"ID":"f10e6bef-1760-4a64-a702-675932d1770c","Type":"ContainerStarted","Data":"bd8de5cb0eff83e5c0bc56a140221cfc955a3826c76ced4b78a8f182115ce5f4"} Dec 27 08:09:23 crc kubenswrapper[4934]: I1227 08:09:23.070717 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-55c9c9fc75-s5tvt" event={"ID":"bb1fd5db-ac54-4e34-9cb3-4caae5f67da8","Type":"ContainerStarted","Data":"9775bf443a0822b6e681ddc018302c60bb6503f841f68ad2ea7e3715cf1d297a"} Dec 27 08:09:23 crc kubenswrapper[4934]: I1227 08:09:23.070761 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-55c9c9fc75-s5tvt" event={"ID":"bb1fd5db-ac54-4e34-9cb3-4caae5f67da8","Type":"ContainerStarted","Data":"998be97c9e5ca0b7d9d3b6a89e314f7a1df936dfcfcc0f57cd931e0ac469b4ae"} Dec 27 08:09:23 crc kubenswrapper[4934]: I1227 08:09:23.072148 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-55c9c9fc75-s5tvt" Dec 27 08:09:23 crc kubenswrapper[4934]: I1227 08:09:23.104888 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-64f664dd6d-jgmlc" event={"ID":"2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43","Type":"ContainerStarted","Data":"d456be41188880d574bc5ece6d790c25db17cd299aed278d76c49136b64e2e03"} Dec 27 08:09:23 crc kubenswrapper[4934]: I1227 08:09:23.104921 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-55c9c9fc75-s5tvt" podStartSLOduration=2.104903063 podStartE2EDuration="2.104903063s" podCreationTimestamp="2025-12-27 08:09:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:09:23.097368406 +0000 UTC m=+1623.917809000" watchObservedRunningTime="2025-12-27 08:09:23.104903063 +0000 UTC m=+1623.925343657" Dec 27 08:09:23 crc kubenswrapper[4934]: I1227 08:09:23.498341 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14b9a667-b34c-4345-8c38-711f74fe9a59" path="/var/lib/kubelet/pods/14b9a667-b34c-4345-8c38-711f74fe9a59/volumes" Dec 27 08:09:23 crc kubenswrapper[4934]: I1227 08:09:23.678384 4934 scope.go:117] "RemoveContainer" containerID="ccfc511f4ea276acde9189dd054526355385fbd2bcc5b07b341eaf9ddbb666ba" Dec 27 08:09:25 crc kubenswrapper[4934]: I1227 08:09:25.136246 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" event={"ID":"f10e6bef-1760-4a64-a702-675932d1770c","Type":"ContainerStarted","Data":"a52d5c84989efa7c86b5d274f6cab205ca791e9eb3a8ecaaba08617e8f80377a"} Dec 27 08:09:25 crc kubenswrapper[4934]: I1227 08:09:25.138074 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:25 crc kubenswrapper[4934]: I1227 08:09:25.140759 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-64f664dd6d-jgmlc" event={"ID":"2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43","Type":"ContainerStarted","Data":"33d39ea805d6701b5e60609e25e82a76d24f93ebae34f4eb8775332a998f86a1"} Dec 27 08:09:25 crc kubenswrapper[4934]: I1227 08:09:25.140895 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:25 crc kubenswrapper[4934]: I1227 08:09:25.162203 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" podStartSLOduration=2.802950622 podStartE2EDuration="4.162178525s" podCreationTimestamp="2025-12-27 08:09:21 +0000 UTC" firstStartedPulling="2025-12-27 08:09:22.625506779 +0000 UTC m=+1623.445947373" lastFinishedPulling="2025-12-27 08:09:23.984734692 +0000 UTC m=+1624.805175276" observedRunningTime="2025-12-27 08:09:25.153866828 +0000 UTC m=+1625.974307422" watchObservedRunningTime="2025-12-27 08:09:25.162178525 +0000 UTC m=+1625.982619109" Dec 27 08:09:25 crc kubenswrapper[4934]: I1227 08:09:25.181208 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-64f664dd6d-jgmlc" podStartSLOduration=2.648891549 podStartE2EDuration="4.181188256s" podCreationTimestamp="2025-12-27 08:09:21 +0000 UTC" firstStartedPulling="2025-12-27 08:09:22.45304433 +0000 UTC m=+1623.273484924" lastFinishedPulling="2025-12-27 08:09:23.985341037 +0000 UTC m=+1624.805781631" observedRunningTime="2025-12-27 08:09:25.173306511 +0000 UTC m=+1625.993747095" watchObservedRunningTime="2025-12-27 08:09:25.181188256 +0000 UTC m=+1626.001628850" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.287221 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b"] Dec 27 08:09:30 crc kubenswrapper[4934]: E1227 08:09:30.288798 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14b9a667-b34c-4345-8c38-711f74fe9a59" containerName="dnsmasq-dns" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.288823 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="14b9a667-b34c-4345-8c38-711f74fe9a59" containerName="dnsmasq-dns" Dec 27 08:09:30 crc kubenswrapper[4934]: E1227 08:09:30.288860 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14b9a667-b34c-4345-8c38-711f74fe9a59" containerName="init" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.288874 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="14b9a667-b34c-4345-8c38-711f74fe9a59" containerName="init" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.289387 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="14b9a667-b34c-4345-8c38-711f74fe9a59" containerName="dnsmasq-dns" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.292478 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.300607 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.300616 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.300745 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-98ddn" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.300786 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.310340 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b"] Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.376665 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3833a830-a283-4755-8909-c5a5c7a218a9-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b\" (UID: \"3833a830-a283-4755-8909-c5a5c7a218a9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.376987 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3833a830-a283-4755-8909-c5a5c7a218a9-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b\" (UID: \"3833a830-a283-4755-8909-c5a5c7a218a9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.377393 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3833a830-a283-4755-8909-c5a5c7a218a9-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b\" (UID: \"3833a830-a283-4755-8909-c5a5c7a218a9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.377597 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nksrg\" (UniqueName: \"kubernetes.io/projected/3833a830-a283-4755-8909-c5a5c7a218a9-kube-api-access-nksrg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b\" (UID: \"3833a830-a283-4755-8909-c5a5c7a218a9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.468628 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:09:30 crc kubenswrapper[4934]: E1227 08:09:30.469280 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.479542 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3833a830-a283-4755-8909-c5a5c7a218a9-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b\" (UID: \"3833a830-a283-4755-8909-c5a5c7a218a9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.479756 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nksrg\" (UniqueName: \"kubernetes.io/projected/3833a830-a283-4755-8909-c5a5c7a218a9-kube-api-access-nksrg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b\" (UID: \"3833a830-a283-4755-8909-c5a5c7a218a9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.484550 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3833a830-a283-4755-8909-c5a5c7a218a9-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b\" (UID: \"3833a830-a283-4755-8909-c5a5c7a218a9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.484782 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3833a830-a283-4755-8909-c5a5c7a218a9-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b\" (UID: \"3833a830-a283-4755-8909-c5a5c7a218a9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.488043 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3833a830-a283-4755-8909-c5a5c7a218a9-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b\" (UID: \"3833a830-a283-4755-8909-c5a5c7a218a9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.488427 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3833a830-a283-4755-8909-c5a5c7a218a9-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b\" (UID: \"3833a830-a283-4755-8909-c5a5c7a218a9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.490153 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3833a830-a283-4755-8909-c5a5c7a218a9-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b\" (UID: \"3833a830-a283-4755-8909-c5a5c7a218a9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.499728 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nksrg\" (UniqueName: \"kubernetes.io/projected/3833a830-a283-4755-8909-c5a5c7a218a9-kube-api-access-nksrg\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b\" (UID: \"3833a830-a283-4755-8909-c5a5c7a218a9\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b" Dec 27 08:09:30 crc kubenswrapper[4934]: I1227 08:09:30.640921 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b" Dec 27 08:09:31 crc kubenswrapper[4934]: I1227 08:09:31.454972 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b"] Dec 27 08:09:32 crc kubenswrapper[4934]: I1227 08:09:32.223654 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b" event={"ID":"3833a830-a283-4755-8909-c5a5c7a218a9","Type":"ContainerStarted","Data":"483b2943d8915aa758b5195ddc660cb0b8c7d9cd9859914da394c2d9187a1fbc"} Dec 27 08:09:33 crc kubenswrapper[4934]: I1227 08:09:33.383691 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-8cb7b99d5-5xbmt" Dec 27 08:09:33 crc kubenswrapper[4934]: I1227 08:09:33.392156 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-64f664dd6d-jgmlc" Dec 27 08:09:33 crc kubenswrapper[4934]: I1227 08:09:33.507180 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-cf699bf96-qsv9c"] Dec 27 08:09:33 crc kubenswrapper[4934]: I1227 08:09:33.507693 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-cf699bf96-qsv9c" podUID="67b18dc3-52fb-4d6c-81de-ec66d3e19e57" containerName="heat-cfnapi" containerID="cri-o://98466e60fc6c1c49235813b1ce6248fb7e70cfdf95fa0e64f6a8ccdac3d08039" gracePeriod=60 Dec 27 08:09:33 crc kubenswrapper[4934]: I1227 08:09:33.524985 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-785dfb69f-wftff"] Dec 27 08:09:33 crc kubenswrapper[4934]: I1227 08:09:33.525335 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-785dfb69f-wftff" podUID="d44e2313-96f9-475a-8adf-f8ba1c9ffb0b" containerName="heat-api" containerID="cri-o://d8ca97c2fc969b024a50a5e3cb3016115c652ed613bdfbefeb1b6bbba9809849" gracePeriod=60 Dec 27 08:09:36 crc kubenswrapper[4934]: I1227 08:09:36.688527 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-785dfb69f-wftff" podUID="d44e2313-96f9-475a-8adf-f8ba1c9ffb0b" containerName="heat-api" probeResult="failure" output="Get \"https://10.217.0.216:8004/healthcheck\": read tcp 10.217.0.2:44798->10.217.0.216:8004: read: connection reset by peer" Dec 27 08:09:36 crc kubenswrapper[4934]: I1227 08:09:36.690675 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-cf699bf96-qsv9c" podUID="67b18dc3-52fb-4d6c-81de-ec66d3e19e57" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.0.217:8000/healthcheck\": read tcp 10.217.0.2:36478->10.217.0.217:8000: read: connection reset by peer" Dec 27 08:09:37 crc kubenswrapper[4934]: I1227 08:09:37.290037 4934 generic.go:334] "Generic (PLEG): container finished" podID="d44e2313-96f9-475a-8adf-f8ba1c9ffb0b" containerID="d8ca97c2fc969b024a50a5e3cb3016115c652ed613bdfbefeb1b6bbba9809849" exitCode=0 Dec 27 08:09:37 crc kubenswrapper[4934]: I1227 08:09:37.290118 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-785dfb69f-wftff" event={"ID":"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b","Type":"ContainerDied","Data":"d8ca97c2fc969b024a50a5e3cb3016115c652ed613bdfbefeb1b6bbba9809849"} Dec 27 08:09:37 crc kubenswrapper[4934]: I1227 08:09:37.292614 4934 generic.go:334] "Generic (PLEG): container finished" podID="85c58dd9-a367-4468-ad44-0f567a89c903" containerID="795f5acf5725079964e7a78f9a2820082b4abd45fe0a050721438cd2d158857d" exitCode=0 Dec 27 08:09:37 crc kubenswrapper[4934]: I1227 08:09:37.292674 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"85c58dd9-a367-4468-ad44-0f567a89c903","Type":"ContainerDied","Data":"795f5acf5725079964e7a78f9a2820082b4abd45fe0a050721438cd2d158857d"} Dec 27 08:09:37 crc kubenswrapper[4934]: I1227 08:09:37.298200 4934 generic.go:334] "Generic (PLEG): container finished" podID="67b18dc3-52fb-4d6c-81de-ec66d3e19e57" containerID="98466e60fc6c1c49235813b1ce6248fb7e70cfdf95fa0e64f6a8ccdac3d08039" exitCode=0 Dec 27 08:09:37 crc kubenswrapper[4934]: I1227 08:09:37.298251 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-cf699bf96-qsv9c" event={"ID":"67b18dc3-52fb-4d6c-81de-ec66d3e19e57","Type":"ContainerDied","Data":"98466e60fc6c1c49235813b1ce6248fb7e70cfdf95fa0e64f6a8ccdac3d08039"} Dec 27 08:09:38 crc kubenswrapper[4934]: I1227 08:09:38.319936 4934 generic.go:334] "Generic (PLEG): container finished" podID="5a3ecad3-f6df-4051-be94-faff5cbc4e1f" containerID="2502034eeae116a80d839154687bc95479ad709015cfb7946d76d6ad6f13679a" exitCode=0 Dec 27 08:09:38 crc kubenswrapper[4934]: I1227 08:09:38.320376 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"5a3ecad3-f6df-4051-be94-faff5cbc4e1f","Type":"ContainerDied","Data":"2502034eeae116a80d839154687bc95479ad709015cfb7946d76d6ad6f13679a"} Dec 27 08:09:40 crc kubenswrapper[4934]: I1227 08:09:40.960638 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-785dfb69f-wftff" podUID="d44e2313-96f9-475a-8adf-f8ba1c9ffb0b" containerName="heat-api" probeResult="failure" output="Get \"https://10.217.0.216:8004/healthcheck\": dial tcp 10.217.0.216:8004: connect: connection refused" Dec 27 08:09:40 crc kubenswrapper[4934]: I1227 08:09:40.968761 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-cf699bf96-qsv9c" podUID="67b18dc3-52fb-4d6c-81de-ec66d3e19e57" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.0.217:8000/healthcheck\": dial tcp 10.217.0.217:8000: connect: connection refused" Dec 27 08:09:41 crc kubenswrapper[4934]: I1227 08:09:41.753821 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-55c9c9fc75-s5tvt" Dec 27 08:09:41 crc kubenswrapper[4934]: I1227 08:09:41.836108 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-6f98b67c6-55vgd"] Dec 27 08:09:41 crc kubenswrapper[4934]: I1227 08:09:41.836397 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-6f98b67c6-55vgd" podUID="1515c241-e28d-49d7-a3f3-790aa082251c" containerName="heat-engine" containerID="cri-o://80405edd7bb455015027bd7125e6267a53d052b855fe164cffc53d0ef0acd2da" gracePeriod=60 Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.293263 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.336786 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-config-data-custom\") pod \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.337161 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rc8z\" (UniqueName: \"kubernetes.io/projected/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-kube-api-access-9rc8z\") pod \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.337210 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-internal-tls-certs\") pod \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.337245 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-public-tls-certs\") pod \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.337302 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-config-data\") pod \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.337338 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-combined-ca-bundle\") pod \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\" (UID: \"67b18dc3-52fb-4d6c-81de-ec66d3e19e57\") " Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.391169 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "67b18dc3-52fb-4d6c-81de-ec66d3e19e57" (UID: "67b18dc3-52fb-4d6c-81de-ec66d3e19e57"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.393971 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-cf699bf96-qsv9c" event={"ID":"67b18dc3-52fb-4d6c-81de-ec66d3e19e57","Type":"ContainerDied","Data":"4f7e9d0fc3bef95eb6260f173ed1572969592456b56d7f09f2545820b169fc35"} Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.394050 4934 scope.go:117] "RemoveContainer" containerID="98466e60fc6c1c49235813b1ce6248fb7e70cfdf95fa0e64f6a8ccdac3d08039" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.394251 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-cf699bf96-qsv9c" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.400993 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-kube-api-access-9rc8z" (OuterVolumeSpecName: "kube-api-access-9rc8z") pod "67b18dc3-52fb-4d6c-81de-ec66d3e19e57" (UID: "67b18dc3-52fb-4d6c-81de-ec66d3e19e57"). InnerVolumeSpecName "kube-api-access-9rc8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.448056 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.448160 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rc8z\" (UniqueName: \"kubernetes.io/projected/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-kube-api-access-9rc8z\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.469127 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:09:42 crc kubenswrapper[4934]: E1227 08:09:42.469674 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.486720 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.550210 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-combined-ca-bundle\") pod \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.550611 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s24vm\" (UniqueName: \"kubernetes.io/projected/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-kube-api-access-s24vm\") pod \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.550790 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-config-data\") pod \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.550968 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-public-tls-certs\") pod \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.551187 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-internal-tls-certs\") pod \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.551298 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-config-data-custom\") pod \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\" (UID: \"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b\") " Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.584097 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-kube-api-access-s24vm" (OuterVolumeSpecName: "kube-api-access-s24vm") pod "d44e2313-96f9-475a-8adf-f8ba1c9ffb0b" (UID: "d44e2313-96f9-475a-8adf-f8ba1c9ffb0b"). InnerVolumeSpecName "kube-api-access-s24vm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.593347 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d44e2313-96f9-475a-8adf-f8ba1c9ffb0b" (UID: "d44e2313-96f9-475a-8adf-f8ba1c9ffb0b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.655736 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s24vm\" (UniqueName: \"kubernetes.io/projected/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-kube-api-access-s24vm\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.655771 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.664442 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67b18dc3-52fb-4d6c-81de-ec66d3e19e57" (UID: "67b18dc3-52fb-4d6c-81de-ec66d3e19e57"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.695646 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-config-data" (OuterVolumeSpecName: "config-data") pod "67b18dc3-52fb-4d6c-81de-ec66d3e19e57" (UID: "67b18dc3-52fb-4d6c-81de-ec66d3e19e57"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.696178 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "67b18dc3-52fb-4d6c-81de-ec66d3e19e57" (UID: "67b18dc3-52fb-4d6c-81de-ec66d3e19e57"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.698185 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d44e2313-96f9-475a-8adf-f8ba1c9ffb0b" (UID: "d44e2313-96f9-475a-8adf-f8ba1c9ffb0b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:09:42 crc kubenswrapper[4934]: E1227 08:09:42.711881 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="80405edd7bb455015027bd7125e6267a53d052b855fe164cffc53d0ef0acd2da" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 27 08:09:42 crc kubenswrapper[4934]: E1227 08:09:42.716492 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="80405edd7bb455015027bd7125e6267a53d052b855fe164cffc53d0ef0acd2da" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 27 08:09:42 crc kubenswrapper[4934]: E1227 08:09:42.719252 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="80405edd7bb455015027bd7125e6267a53d052b855fe164cffc53d0ef0acd2da" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 27 08:09:42 crc kubenswrapper[4934]: E1227 08:09:42.719416 4934 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-6f98b67c6-55vgd" podUID="1515c241-e28d-49d7-a3f3-790aa082251c" containerName="heat-engine" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.722565 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-config-data" (OuterVolumeSpecName: "config-data") pod "d44e2313-96f9-475a-8adf-f8ba1c9ffb0b" (UID: "d44e2313-96f9-475a-8adf-f8ba1c9ffb0b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.725270 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d44e2313-96f9-475a-8adf-f8ba1c9ffb0b" (UID: "d44e2313-96f9-475a-8adf-f8ba1c9ffb0b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.727192 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "67b18dc3-52fb-4d6c-81de-ec66d3e19e57" (UID: "67b18dc3-52fb-4d6c-81de-ec66d3e19e57"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.757666 4934 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.757708 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.757717 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.757725 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.757735 4934 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.757744 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.757753 4934 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/67b18dc3-52fb-4d6c-81de-ec66d3e19e57-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.761442 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d44e2313-96f9-475a-8adf-f8ba1c9ffb0b" (UID: "d44e2313-96f9-475a-8adf-f8ba1c9ffb0b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:09:42 crc kubenswrapper[4934]: I1227 08:09:42.860236 4934 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:43 crc kubenswrapper[4934]: I1227 08:09:43.037666 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-cf699bf96-qsv9c"] Dec 27 08:09:43 crc kubenswrapper[4934]: I1227 08:09:43.048365 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-cf699bf96-qsv9c"] Dec 27 08:09:43 crc kubenswrapper[4934]: I1227 08:09:43.408429 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-785dfb69f-wftff" event={"ID":"d44e2313-96f9-475a-8adf-f8ba1c9ffb0b","Type":"ContainerDied","Data":"0599de68b5101a784656b736a194b29dde57feeed87b0890540c39580210147f"} Dec 27 08:09:43 crc kubenswrapper[4934]: I1227 08:09:43.408480 4934 scope.go:117] "RemoveContainer" containerID="d8ca97c2fc969b024a50a5e3cb3016115c652ed613bdfbefeb1b6bbba9809849" Dec 27 08:09:43 crc kubenswrapper[4934]: I1227 08:09:43.408473 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-785dfb69f-wftff" Dec 27 08:09:43 crc kubenswrapper[4934]: I1227 08:09:43.410773 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"5a3ecad3-f6df-4051-be94-faff5cbc4e1f","Type":"ContainerStarted","Data":"dc0cf12b7c1864cad3d9a894f473c88495cba693993f22c9ca24e35187849cdb"} Dec 27 08:09:43 crc kubenswrapper[4934]: I1227 08:09:43.410987 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-2" Dec 27 08:09:43 crc kubenswrapper[4934]: I1227 08:09:43.426568 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"85c58dd9-a367-4468-ad44-0f567a89c903","Type":"ContainerStarted","Data":"ababf99f0d9d5a81abe7c2b33e85c0abddc7d5b9c09ad5133a14046080986014"} Dec 27 08:09:43 crc kubenswrapper[4934]: I1227 08:09:43.427493 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:09:43 crc kubenswrapper[4934]: I1227 08:09:43.433737 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b" event={"ID":"3833a830-a283-4755-8909-c5a5c7a218a9","Type":"ContainerStarted","Data":"aa4cd0d1252bd3173a37c2c38e30f12ed24a6d08e7a9622511a602673fab989f"} Dec 27 08:09:43 crc kubenswrapper[4934]: I1227 08:09:43.486022 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67b18dc3-52fb-4d6c-81de-ec66d3e19e57" path="/var/lib/kubelet/pods/67b18dc3-52fb-4d6c-81de-ec66d3e19e57/volumes" Dec 27 08:09:43 crc kubenswrapper[4934]: I1227 08:09:43.507012 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-2" podStartSLOduration=41.506995017 podStartE2EDuration="41.506995017s" podCreationTimestamp="2025-12-27 08:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:09:43.483573716 +0000 UTC m=+1644.304014310" watchObservedRunningTime="2025-12-27 08:09:43.506995017 +0000 UTC m=+1644.327435611" Dec 27 08:09:43 crc kubenswrapper[4934]: I1227 08:09:43.551999 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=42.551979413 podStartE2EDuration="42.551979413s" podCreationTimestamp="2025-12-27 08:09:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:09:43.534639673 +0000 UTC m=+1644.355080287" watchObservedRunningTime="2025-12-27 08:09:43.551979413 +0000 UTC m=+1644.372420007" Dec 27 08:09:43 crc kubenswrapper[4934]: I1227 08:09:43.576177 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-785dfb69f-wftff"] Dec 27 08:09:43 crc kubenswrapper[4934]: I1227 08:09:43.616146 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-785dfb69f-wftff"] Dec 27 08:09:43 crc kubenswrapper[4934]: I1227 08:09:43.632431 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b" podStartSLOduration=3.059619449 podStartE2EDuration="13.632409259s" podCreationTimestamp="2025-12-27 08:09:30 +0000 UTC" firstStartedPulling="2025-12-27 08:09:31.46756192 +0000 UTC m=+1632.288002504" lastFinishedPulling="2025-12-27 08:09:42.04035172 +0000 UTC m=+1642.860792314" observedRunningTime="2025-12-27 08:09:43.604433405 +0000 UTC m=+1644.424873999" watchObservedRunningTime="2025-12-27 08:09:43.632409259 +0000 UTC m=+1644.452849853" Dec 27 08:09:45 crc kubenswrapper[4934]: I1227 08:09:45.486306 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d44e2313-96f9-475a-8adf-f8ba1c9ffb0b" path="/var/lib/kubelet/pods/d44e2313-96f9-475a-8adf-f8ba1c9ffb0b/volumes" Dec 27 08:09:46 crc kubenswrapper[4934]: I1227 08:09:46.837941 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 27 08:09:47 crc kubenswrapper[4934]: I1227 08:09:47.810148 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-d4fxd"] Dec 27 08:09:47 crc kubenswrapper[4934]: I1227 08:09:47.821523 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-d4fxd"] Dec 27 08:09:47 crc kubenswrapper[4934]: I1227 08:09:47.890609 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-wgpjv"] Dec 27 08:09:47 crc kubenswrapper[4934]: E1227 08:09:47.891142 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67b18dc3-52fb-4d6c-81de-ec66d3e19e57" containerName="heat-cfnapi" Dec 27 08:09:47 crc kubenswrapper[4934]: I1227 08:09:47.891154 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="67b18dc3-52fb-4d6c-81de-ec66d3e19e57" containerName="heat-cfnapi" Dec 27 08:09:47 crc kubenswrapper[4934]: E1227 08:09:47.891200 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d44e2313-96f9-475a-8adf-f8ba1c9ffb0b" containerName="heat-api" Dec 27 08:09:47 crc kubenswrapper[4934]: I1227 08:09:47.891206 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d44e2313-96f9-475a-8adf-f8ba1c9ffb0b" containerName="heat-api" Dec 27 08:09:47 crc kubenswrapper[4934]: I1227 08:09:47.891439 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="d44e2313-96f9-475a-8adf-f8ba1c9ffb0b" containerName="heat-api" Dec 27 08:09:47 crc kubenswrapper[4934]: I1227 08:09:47.891459 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="67b18dc3-52fb-4d6c-81de-ec66d3e19e57" containerName="heat-cfnapi" Dec 27 08:09:47 crc kubenswrapper[4934]: I1227 08:09:47.892254 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-wgpjv" Dec 27 08:09:47 crc kubenswrapper[4934]: I1227 08:09:47.908428 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-wgpjv"] Dec 27 08:09:47 crc kubenswrapper[4934]: I1227 08:09:47.981805 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f05689e-1351-435f-a267-f0c6508f1f4a-config-data\") pod \"aodh-db-sync-wgpjv\" (UID: \"4f05689e-1351-435f-a267-f0c6508f1f4a\") " pod="openstack/aodh-db-sync-wgpjv" Dec 27 08:09:47 crc kubenswrapper[4934]: I1227 08:09:47.981876 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f05689e-1351-435f-a267-f0c6508f1f4a-scripts\") pod \"aodh-db-sync-wgpjv\" (UID: \"4f05689e-1351-435f-a267-f0c6508f1f4a\") " pod="openstack/aodh-db-sync-wgpjv" Dec 27 08:09:47 crc kubenswrapper[4934]: I1227 08:09:47.982024 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f05689e-1351-435f-a267-f0c6508f1f4a-combined-ca-bundle\") pod \"aodh-db-sync-wgpjv\" (UID: \"4f05689e-1351-435f-a267-f0c6508f1f4a\") " pod="openstack/aodh-db-sync-wgpjv" Dec 27 08:09:47 crc kubenswrapper[4934]: I1227 08:09:47.982150 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pv2j9\" (UniqueName: \"kubernetes.io/projected/4f05689e-1351-435f-a267-f0c6508f1f4a-kube-api-access-pv2j9\") pod \"aodh-db-sync-wgpjv\" (UID: \"4f05689e-1351-435f-a267-f0c6508f1f4a\") " pod="openstack/aodh-db-sync-wgpjv" Dec 27 08:09:48 crc kubenswrapper[4934]: I1227 08:09:48.083620 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pv2j9\" (UniqueName: \"kubernetes.io/projected/4f05689e-1351-435f-a267-f0c6508f1f4a-kube-api-access-pv2j9\") pod \"aodh-db-sync-wgpjv\" (UID: \"4f05689e-1351-435f-a267-f0c6508f1f4a\") " pod="openstack/aodh-db-sync-wgpjv" Dec 27 08:09:48 crc kubenswrapper[4934]: I1227 08:09:48.083766 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f05689e-1351-435f-a267-f0c6508f1f4a-config-data\") pod \"aodh-db-sync-wgpjv\" (UID: \"4f05689e-1351-435f-a267-f0c6508f1f4a\") " pod="openstack/aodh-db-sync-wgpjv" Dec 27 08:09:48 crc kubenswrapper[4934]: I1227 08:09:48.083802 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f05689e-1351-435f-a267-f0c6508f1f4a-scripts\") pod \"aodh-db-sync-wgpjv\" (UID: \"4f05689e-1351-435f-a267-f0c6508f1f4a\") " pod="openstack/aodh-db-sync-wgpjv" Dec 27 08:09:48 crc kubenswrapper[4934]: I1227 08:09:48.084606 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f05689e-1351-435f-a267-f0c6508f1f4a-combined-ca-bundle\") pod \"aodh-db-sync-wgpjv\" (UID: \"4f05689e-1351-435f-a267-f0c6508f1f4a\") " pod="openstack/aodh-db-sync-wgpjv" Dec 27 08:09:48 crc kubenswrapper[4934]: I1227 08:09:48.088964 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f05689e-1351-435f-a267-f0c6508f1f4a-scripts\") pod \"aodh-db-sync-wgpjv\" (UID: \"4f05689e-1351-435f-a267-f0c6508f1f4a\") " pod="openstack/aodh-db-sync-wgpjv" Dec 27 08:09:48 crc kubenswrapper[4934]: I1227 08:09:48.089552 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f05689e-1351-435f-a267-f0c6508f1f4a-config-data\") pod \"aodh-db-sync-wgpjv\" (UID: \"4f05689e-1351-435f-a267-f0c6508f1f4a\") " pod="openstack/aodh-db-sync-wgpjv" Dec 27 08:09:48 crc kubenswrapper[4934]: I1227 08:09:48.089754 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f05689e-1351-435f-a267-f0c6508f1f4a-combined-ca-bundle\") pod \"aodh-db-sync-wgpjv\" (UID: \"4f05689e-1351-435f-a267-f0c6508f1f4a\") " pod="openstack/aodh-db-sync-wgpjv" Dec 27 08:09:48 crc kubenswrapper[4934]: I1227 08:09:48.106774 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pv2j9\" (UniqueName: \"kubernetes.io/projected/4f05689e-1351-435f-a267-f0c6508f1f4a-kube-api-access-pv2j9\") pod \"aodh-db-sync-wgpjv\" (UID: \"4f05689e-1351-435f-a267-f0c6508f1f4a\") " pod="openstack/aodh-db-sync-wgpjv" Dec 27 08:09:48 crc kubenswrapper[4934]: I1227 08:09:48.220332 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-wgpjv" Dec 27 08:09:48 crc kubenswrapper[4934]: W1227 08:09:48.774231 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f05689e_1351_435f_a267_f0c6508f1f4a.slice/crio-b158a12ea7620bcef0662864b97c9cb7c14dff3e64b540570c96ad15460f35b8 WatchSource:0}: Error finding container b158a12ea7620bcef0662864b97c9cb7c14dff3e64b540570c96ad15460f35b8: Status 404 returned error can't find the container with id b158a12ea7620bcef0662864b97c9cb7c14dff3e64b540570c96ad15460f35b8 Dec 27 08:09:48 crc kubenswrapper[4934]: I1227 08:09:48.778332 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-wgpjv"] Dec 27 08:09:49 crc kubenswrapper[4934]: I1227 08:09:49.493965 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e83bd090-ad9e-4b6a-b72f-b6bdab47127a" path="/var/lib/kubelet/pods/e83bd090-ad9e-4b6a-b72f-b6bdab47127a/volumes" Dec 27 08:09:49 crc kubenswrapper[4934]: I1227 08:09:49.530555 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-wgpjv" event={"ID":"4f05689e-1351-435f-a267-f0c6508f1f4a","Type":"ContainerStarted","Data":"b158a12ea7620bcef0662864b97c9cb7c14dff3e64b540570c96ad15460f35b8"} Dec 27 08:09:52 crc kubenswrapper[4934]: I1227 08:09:52.329468 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="85c58dd9-a367-4468-ad44-0f567a89c903" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.11:5671: connect: connection refused" Dec 27 08:09:52 crc kubenswrapper[4934]: E1227 08:09:52.708594 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="80405edd7bb455015027bd7125e6267a53d052b855fe164cffc53d0ef0acd2da" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 27 08:09:52 crc kubenswrapper[4934]: E1227 08:09:52.711282 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="80405edd7bb455015027bd7125e6267a53d052b855fe164cffc53d0ef0acd2da" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 27 08:09:52 crc kubenswrapper[4934]: E1227 08:09:52.712892 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="80405edd7bb455015027bd7125e6267a53d052b855fe164cffc53d0ef0acd2da" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 27 08:09:52 crc kubenswrapper[4934]: E1227 08:09:52.712951 4934 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-6f98b67c6-55vgd" podUID="1515c241-e28d-49d7-a3f3-790aa082251c" containerName="heat-engine" Dec 27 08:09:53 crc kubenswrapper[4934]: I1227 08:09:53.325012 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="5a3ecad3-f6df-4051-be94-faff5cbc4e1f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.12:5671: connect: connection refused" Dec 27 08:09:55 crc kubenswrapper[4934]: I1227 08:09:55.468838 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:09:55 crc kubenswrapper[4934]: E1227 08:09:55.469784 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:09:55 crc kubenswrapper[4934]: I1227 08:09:55.639991 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-wgpjv" event={"ID":"4f05689e-1351-435f-a267-f0c6508f1f4a","Type":"ContainerStarted","Data":"d4029fef32d31d19e82de9403dc39d655329d6235f402b4de52d8250f613f1f0"} Dec 27 08:09:56 crc kubenswrapper[4934]: I1227 08:09:56.656328 4934 generic.go:334] "Generic (PLEG): container finished" podID="3833a830-a283-4755-8909-c5a5c7a218a9" containerID="aa4cd0d1252bd3173a37c2c38e30f12ed24a6d08e7a9622511a602673fab989f" exitCode=0 Dec 27 08:09:56 crc kubenswrapper[4934]: I1227 08:09:56.656465 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b" event={"ID":"3833a830-a283-4755-8909-c5a5c7a218a9","Type":"ContainerDied","Data":"aa4cd0d1252bd3173a37c2c38e30f12ed24a6d08e7a9622511a602673fab989f"} Dec 27 08:09:56 crc kubenswrapper[4934]: I1227 08:09:56.764690 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-wgpjv" podStartSLOduration=3.697812412 podStartE2EDuration="9.76467124s" podCreationTimestamp="2025-12-27 08:09:47 +0000 UTC" firstStartedPulling="2025-12-27 08:09:48.78121324 +0000 UTC m=+1649.601653834" lastFinishedPulling="2025-12-27 08:09:54.848072068 +0000 UTC m=+1655.668512662" observedRunningTime="2025-12-27 08:09:55.663180541 +0000 UTC m=+1656.483621125" watchObservedRunningTime="2025-12-27 08:09:56.76467124 +0000 UTC m=+1657.585111834" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.255996 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.344830 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3833a830-a283-4755-8909-c5a5c7a218a9-ssh-key\") pod \"3833a830-a283-4755-8909-c5a5c7a218a9\" (UID: \"3833a830-a283-4755-8909-c5a5c7a218a9\") " Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.345116 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3833a830-a283-4755-8909-c5a5c7a218a9-inventory\") pod \"3833a830-a283-4755-8909-c5a5c7a218a9\" (UID: \"3833a830-a283-4755-8909-c5a5c7a218a9\") " Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.345359 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3833a830-a283-4755-8909-c5a5c7a218a9-repo-setup-combined-ca-bundle\") pod \"3833a830-a283-4755-8909-c5a5c7a218a9\" (UID: \"3833a830-a283-4755-8909-c5a5c7a218a9\") " Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.345446 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nksrg\" (UniqueName: \"kubernetes.io/projected/3833a830-a283-4755-8909-c5a5c7a218a9-kube-api-access-nksrg\") pod \"3833a830-a283-4755-8909-c5a5c7a218a9\" (UID: \"3833a830-a283-4755-8909-c5a5c7a218a9\") " Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.350944 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3833a830-a283-4755-8909-c5a5c7a218a9-kube-api-access-nksrg" (OuterVolumeSpecName: "kube-api-access-nksrg") pod "3833a830-a283-4755-8909-c5a5c7a218a9" (UID: "3833a830-a283-4755-8909-c5a5c7a218a9"). InnerVolumeSpecName "kube-api-access-nksrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.351252 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3833a830-a283-4755-8909-c5a5c7a218a9-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "3833a830-a283-4755-8909-c5a5c7a218a9" (UID: "3833a830-a283-4755-8909-c5a5c7a218a9"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.385673 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3833a830-a283-4755-8909-c5a5c7a218a9-inventory" (OuterVolumeSpecName: "inventory") pod "3833a830-a283-4755-8909-c5a5c7a218a9" (UID: "3833a830-a283-4755-8909-c5a5c7a218a9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.392256 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3833a830-a283-4755-8909-c5a5c7a218a9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3833a830-a283-4755-8909-c5a5c7a218a9" (UID: "3833a830-a283-4755-8909-c5a5c7a218a9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.449713 4934 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3833a830-a283-4755-8909-c5a5c7a218a9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.449742 4934 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3833a830-a283-4755-8909-c5a5c7a218a9-inventory\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.449751 4934 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3833a830-a283-4755-8909-c5a5c7a218a9-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.449761 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nksrg\" (UniqueName: \"kubernetes.io/projected/3833a830-a283-4755-8909-c5a5c7a218a9-kube-api-access-nksrg\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.691147 4934 generic.go:334] "Generic (PLEG): container finished" podID="1515c241-e28d-49d7-a3f3-790aa082251c" containerID="80405edd7bb455015027bd7125e6267a53d052b855fe164cffc53d0ef0acd2da" exitCode=0 Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.691221 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6f98b67c6-55vgd" event={"ID":"1515c241-e28d-49d7-a3f3-790aa082251c","Type":"ContainerDied","Data":"80405edd7bb455015027bd7125e6267a53d052b855fe164cffc53d0ef0acd2da"} Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.691252 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6f98b67c6-55vgd" event={"ID":"1515c241-e28d-49d7-a3f3-790aa082251c","Type":"ContainerDied","Data":"6d6cdc0c6658bbfb32e7990958b1a7e92df2cc9548f50d710ece359edd8cf5a4"} Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.691265 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d6cdc0c6658bbfb32e7990958b1a7e92df2cc9548f50d710ece359edd8cf5a4" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.692988 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b" event={"ID":"3833a830-a283-4755-8909-c5a5c7a218a9","Type":"ContainerDied","Data":"483b2943d8915aa758b5195ddc660cb0b8c7d9cd9859914da394c2d9187a1fbc"} Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.693014 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="483b2943d8915aa758b5195ddc660cb0b8c7d9cd9859914da394c2d9187a1fbc" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.693105 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.695208 4934 generic.go:334] "Generic (PLEG): container finished" podID="4f05689e-1351-435f-a267-f0c6508f1f4a" containerID="d4029fef32d31d19e82de9403dc39d655329d6235f402b4de52d8250f613f1f0" exitCode=0 Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.695271 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-wgpjv" event={"ID":"4f05689e-1351-435f-a267-f0c6508f1f4a","Type":"ContainerDied","Data":"d4029fef32d31d19e82de9403dc39d655329d6235f402b4de52d8250f613f1f0"} Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.767074 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6f98b67c6-55vgd" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.830566 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-mrdpt"] Dec 27 08:09:58 crc kubenswrapper[4934]: E1227 08:09:58.831165 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3833a830-a283-4755-8909-c5a5c7a218a9" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.831182 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="3833a830-a283-4755-8909-c5a5c7a218a9" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 27 08:09:58 crc kubenswrapper[4934]: E1227 08:09:58.831222 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1515c241-e28d-49d7-a3f3-790aa082251c" containerName="heat-engine" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.831229 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="1515c241-e28d-49d7-a3f3-790aa082251c" containerName="heat-engine" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.831454 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="3833a830-a283-4755-8909-c5a5c7a218a9" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.831483 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="1515c241-e28d-49d7-a3f3-790aa082251c" containerName="heat-engine" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.832314 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mrdpt" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.835898 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.835998 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.836050 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-98ddn" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.836180 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.844699 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-mrdpt"] Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.860066 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1515c241-e28d-49d7-a3f3-790aa082251c-combined-ca-bundle\") pod \"1515c241-e28d-49d7-a3f3-790aa082251c\" (UID: \"1515c241-e28d-49d7-a3f3-790aa082251c\") " Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.860193 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1515c241-e28d-49d7-a3f3-790aa082251c-config-data-custom\") pod \"1515c241-e28d-49d7-a3f3-790aa082251c\" (UID: \"1515c241-e28d-49d7-a3f3-790aa082251c\") " Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.860336 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8gtw\" (UniqueName: \"kubernetes.io/projected/1515c241-e28d-49d7-a3f3-790aa082251c-kube-api-access-j8gtw\") pod \"1515c241-e28d-49d7-a3f3-790aa082251c\" (UID: \"1515c241-e28d-49d7-a3f3-790aa082251c\") " Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.860369 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1515c241-e28d-49d7-a3f3-790aa082251c-config-data\") pod \"1515c241-e28d-49d7-a3f3-790aa082251c\" (UID: \"1515c241-e28d-49d7-a3f3-790aa082251c\") " Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.872389 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1515c241-e28d-49d7-a3f3-790aa082251c-kube-api-access-j8gtw" (OuterVolumeSpecName: "kube-api-access-j8gtw") pod "1515c241-e28d-49d7-a3f3-790aa082251c" (UID: "1515c241-e28d-49d7-a3f3-790aa082251c"). InnerVolumeSpecName "kube-api-access-j8gtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.874884 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1515c241-e28d-49d7-a3f3-790aa082251c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1515c241-e28d-49d7-a3f3-790aa082251c" (UID: "1515c241-e28d-49d7-a3f3-790aa082251c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.904838 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1515c241-e28d-49d7-a3f3-790aa082251c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1515c241-e28d-49d7-a3f3-790aa082251c" (UID: "1515c241-e28d-49d7-a3f3-790aa082251c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.931274 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1515c241-e28d-49d7-a3f3-790aa082251c-config-data" (OuterVolumeSpecName: "config-data") pod "1515c241-e28d-49d7-a3f3-790aa082251c" (UID: "1515c241-e28d-49d7-a3f3-790aa082251c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.963114 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61f9c149-e7a9-4ca5-b708-aa9274ee74af-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mrdpt\" (UID: \"61f9c149-e7a9-4ca5-b708-aa9274ee74af\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mrdpt" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.963719 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/61f9c149-e7a9-4ca5-b708-aa9274ee74af-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mrdpt\" (UID: \"61f9c149-e7a9-4ca5-b708-aa9274ee74af\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mrdpt" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.963863 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f66kk\" (UniqueName: \"kubernetes.io/projected/61f9c149-e7a9-4ca5-b708-aa9274ee74af-kube-api-access-f66kk\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mrdpt\" (UID: \"61f9c149-e7a9-4ca5-b708-aa9274ee74af\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mrdpt" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.964166 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1515c241-e28d-49d7-a3f3-790aa082251c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.964231 4934 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1515c241-e28d-49d7-a3f3-790aa082251c-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.964296 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8gtw\" (UniqueName: \"kubernetes.io/projected/1515c241-e28d-49d7-a3f3-790aa082251c-kube-api-access-j8gtw\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:58 crc kubenswrapper[4934]: I1227 08:09:58.964352 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1515c241-e28d-49d7-a3f3-790aa082251c-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:09:59 crc kubenswrapper[4934]: I1227 08:09:59.067505 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61f9c149-e7a9-4ca5-b708-aa9274ee74af-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mrdpt\" (UID: \"61f9c149-e7a9-4ca5-b708-aa9274ee74af\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mrdpt" Dec 27 08:09:59 crc kubenswrapper[4934]: I1227 08:09:59.067863 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/61f9c149-e7a9-4ca5-b708-aa9274ee74af-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mrdpt\" (UID: \"61f9c149-e7a9-4ca5-b708-aa9274ee74af\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mrdpt" Dec 27 08:09:59 crc kubenswrapper[4934]: I1227 08:09:59.068235 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f66kk\" (UniqueName: \"kubernetes.io/projected/61f9c149-e7a9-4ca5-b708-aa9274ee74af-kube-api-access-f66kk\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mrdpt\" (UID: \"61f9c149-e7a9-4ca5-b708-aa9274ee74af\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mrdpt" Dec 27 08:09:59 crc kubenswrapper[4934]: I1227 08:09:59.072228 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/61f9c149-e7a9-4ca5-b708-aa9274ee74af-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mrdpt\" (UID: \"61f9c149-e7a9-4ca5-b708-aa9274ee74af\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mrdpt" Dec 27 08:09:59 crc kubenswrapper[4934]: I1227 08:09:59.074758 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61f9c149-e7a9-4ca5-b708-aa9274ee74af-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mrdpt\" (UID: \"61f9c149-e7a9-4ca5-b708-aa9274ee74af\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mrdpt" Dec 27 08:09:59 crc kubenswrapper[4934]: I1227 08:09:59.098807 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f66kk\" (UniqueName: \"kubernetes.io/projected/61f9c149-e7a9-4ca5-b708-aa9274ee74af-kube-api-access-f66kk\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-mrdpt\" (UID: \"61f9c149-e7a9-4ca5-b708-aa9274ee74af\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mrdpt" Dec 27 08:09:59 crc kubenswrapper[4934]: I1227 08:09:59.169620 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mrdpt" Dec 27 08:09:59 crc kubenswrapper[4934]: I1227 08:09:59.710200 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6f98b67c6-55vgd" Dec 27 08:09:59 crc kubenswrapper[4934]: I1227 08:09:59.750987 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-6f98b67c6-55vgd"] Dec 27 08:09:59 crc kubenswrapper[4934]: I1227 08:09:59.773676 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-6f98b67c6-55vgd"] Dec 27 08:09:59 crc kubenswrapper[4934]: I1227 08:09:59.869268 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-mrdpt"] Dec 27 08:10:00 crc kubenswrapper[4934]: I1227 08:10:00.237437 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-wgpjv" Dec 27 08:10:00 crc kubenswrapper[4934]: I1227 08:10:00.303714 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pv2j9\" (UniqueName: \"kubernetes.io/projected/4f05689e-1351-435f-a267-f0c6508f1f4a-kube-api-access-pv2j9\") pod \"4f05689e-1351-435f-a267-f0c6508f1f4a\" (UID: \"4f05689e-1351-435f-a267-f0c6508f1f4a\") " Dec 27 08:10:00 crc kubenswrapper[4934]: I1227 08:10:00.303862 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f05689e-1351-435f-a267-f0c6508f1f4a-scripts\") pod \"4f05689e-1351-435f-a267-f0c6508f1f4a\" (UID: \"4f05689e-1351-435f-a267-f0c6508f1f4a\") " Dec 27 08:10:00 crc kubenswrapper[4934]: I1227 08:10:00.303894 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f05689e-1351-435f-a267-f0c6508f1f4a-config-data\") pod \"4f05689e-1351-435f-a267-f0c6508f1f4a\" (UID: \"4f05689e-1351-435f-a267-f0c6508f1f4a\") " Dec 27 08:10:00 crc kubenswrapper[4934]: I1227 08:10:00.304183 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f05689e-1351-435f-a267-f0c6508f1f4a-combined-ca-bundle\") pod \"4f05689e-1351-435f-a267-f0c6508f1f4a\" (UID: \"4f05689e-1351-435f-a267-f0c6508f1f4a\") " Dec 27 08:10:00 crc kubenswrapper[4934]: I1227 08:10:00.310855 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f05689e-1351-435f-a267-f0c6508f1f4a-kube-api-access-pv2j9" (OuterVolumeSpecName: "kube-api-access-pv2j9") pod "4f05689e-1351-435f-a267-f0c6508f1f4a" (UID: "4f05689e-1351-435f-a267-f0c6508f1f4a"). InnerVolumeSpecName "kube-api-access-pv2j9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:10:00 crc kubenswrapper[4934]: I1227 08:10:00.317514 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f05689e-1351-435f-a267-f0c6508f1f4a-scripts" (OuterVolumeSpecName: "scripts") pod "4f05689e-1351-435f-a267-f0c6508f1f4a" (UID: "4f05689e-1351-435f-a267-f0c6508f1f4a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:10:00 crc kubenswrapper[4934]: I1227 08:10:00.339879 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f05689e-1351-435f-a267-f0c6508f1f4a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f05689e-1351-435f-a267-f0c6508f1f4a" (UID: "4f05689e-1351-435f-a267-f0c6508f1f4a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:10:00 crc kubenswrapper[4934]: I1227 08:10:00.346505 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f05689e-1351-435f-a267-f0c6508f1f4a-config-data" (OuterVolumeSpecName: "config-data") pod "4f05689e-1351-435f-a267-f0c6508f1f4a" (UID: "4f05689e-1351-435f-a267-f0c6508f1f4a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:10:00 crc kubenswrapper[4934]: I1227 08:10:00.408399 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f05689e-1351-435f-a267-f0c6508f1f4a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:00 crc kubenswrapper[4934]: I1227 08:10:00.408436 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pv2j9\" (UniqueName: \"kubernetes.io/projected/4f05689e-1351-435f-a267-f0c6508f1f4a-kube-api-access-pv2j9\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:00 crc kubenswrapper[4934]: I1227 08:10:00.408451 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f05689e-1351-435f-a267-f0c6508f1f4a-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:00 crc kubenswrapper[4934]: I1227 08:10:00.408462 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f05689e-1351-435f-a267-f0c6508f1f4a-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:00 crc kubenswrapper[4934]: I1227 08:10:00.743511 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mrdpt" event={"ID":"61f9c149-e7a9-4ca5-b708-aa9274ee74af","Type":"ContainerStarted","Data":"0980cced3e2da0bae985f12b4190cebac64e6d17603abd9613b0dfa9c44eaa9d"} Dec 27 08:10:00 crc kubenswrapper[4934]: I1227 08:10:00.743768 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mrdpt" event={"ID":"61f9c149-e7a9-4ca5-b708-aa9274ee74af","Type":"ContainerStarted","Data":"f1a7159e62e960762d4ae6c1ee01b5670c29da309af80b6dbd9e9b6f53a368e0"} Dec 27 08:10:00 crc kubenswrapper[4934]: I1227 08:10:00.770926 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-wgpjv" event={"ID":"4f05689e-1351-435f-a267-f0c6508f1f4a","Type":"ContainerDied","Data":"b158a12ea7620bcef0662864b97c9cb7c14dff3e64b540570c96ad15460f35b8"} Dec 27 08:10:00 crc kubenswrapper[4934]: I1227 08:10:00.771150 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b158a12ea7620bcef0662864b97c9cb7c14dff3e64b540570c96ad15460f35b8" Dec 27 08:10:00 crc kubenswrapper[4934]: I1227 08:10:00.771353 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-wgpjv" Dec 27 08:10:00 crc kubenswrapper[4934]: I1227 08:10:00.789615 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mrdpt" podStartSLOduration=2.320610363 podStartE2EDuration="2.789598968s" podCreationTimestamp="2025-12-27 08:09:58 +0000 UTC" firstStartedPulling="2025-12-27 08:09:59.925336126 +0000 UTC m=+1660.745776720" lastFinishedPulling="2025-12-27 08:10:00.394324731 +0000 UTC m=+1661.214765325" observedRunningTime="2025-12-27 08:10:00.775459147 +0000 UTC m=+1661.595899741" watchObservedRunningTime="2025-12-27 08:10:00.789598968 +0000 UTC m=+1661.610039562" Dec 27 08:10:01 crc kubenswrapper[4934]: I1227 08:10:01.495277 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1515c241-e28d-49d7-a3f3-790aa082251c" path="/var/lib/kubelet/pods/1515c241-e28d-49d7-a3f3-790aa082251c/volumes" Dec 27 08:10:02 crc kubenswrapper[4934]: I1227 08:10:02.328437 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 27 08:10:02 crc kubenswrapper[4934]: I1227 08:10:02.907184 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 27 08:10:02 crc kubenswrapper[4934]: I1227 08:10:02.907468 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="e4cf02a3-796a-450f-a150-5ae72e0f6f7d" containerName="aodh-api" containerID="cri-o://57c30729eec9072ee65d52f9e8e83f6cd9a51139aa6062e2adfb001dfc2bb00e" gracePeriod=30 Dec 27 08:10:02 crc kubenswrapper[4934]: I1227 08:10:02.907570 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="e4cf02a3-796a-450f-a150-5ae72e0f6f7d" containerName="aodh-evaluator" containerID="cri-o://c715a325324352bc15756616883cddc8ed263dbac45506fef4fec90a2dcf9bf2" gracePeriod=30 Dec 27 08:10:02 crc kubenswrapper[4934]: I1227 08:10:02.907663 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="e4cf02a3-796a-450f-a150-5ae72e0f6f7d" containerName="aodh-notifier" containerID="cri-o://95df1e1ab27861114324155498743dc1691daf87b8293aef64c5b5f124040970" gracePeriod=30 Dec 27 08:10:02 crc kubenswrapper[4934]: I1227 08:10:02.907963 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="e4cf02a3-796a-450f-a150-5ae72e0f6f7d" containerName="aodh-listener" containerID="cri-o://4e22b411703fbd8da59ac24c7542b39c88cde422839e82708b596306477ee2c0" gracePeriod=30 Dec 27 08:10:03 crc kubenswrapper[4934]: I1227 08:10:03.323134 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-2" Dec 27 08:10:03 crc kubenswrapper[4934]: I1227 08:10:03.386247 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-1"] Dec 27 08:10:03 crc kubenswrapper[4934]: I1227 08:10:03.810972 4934 generic.go:334] "Generic (PLEG): container finished" podID="61f9c149-e7a9-4ca5-b708-aa9274ee74af" containerID="0980cced3e2da0bae985f12b4190cebac64e6d17603abd9613b0dfa9c44eaa9d" exitCode=0 Dec 27 08:10:03 crc kubenswrapper[4934]: I1227 08:10:03.811055 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mrdpt" event={"ID":"61f9c149-e7a9-4ca5-b708-aa9274ee74af","Type":"ContainerDied","Data":"0980cced3e2da0bae985f12b4190cebac64e6d17603abd9613b0dfa9c44eaa9d"} Dec 27 08:10:03 crc kubenswrapper[4934]: I1227 08:10:03.815795 4934 generic.go:334] "Generic (PLEG): container finished" podID="e4cf02a3-796a-450f-a150-5ae72e0f6f7d" containerID="c715a325324352bc15756616883cddc8ed263dbac45506fef4fec90a2dcf9bf2" exitCode=0 Dec 27 08:10:03 crc kubenswrapper[4934]: I1227 08:10:03.815827 4934 generic.go:334] "Generic (PLEG): container finished" podID="e4cf02a3-796a-450f-a150-5ae72e0f6f7d" containerID="57c30729eec9072ee65d52f9e8e83f6cd9a51139aa6062e2adfb001dfc2bb00e" exitCode=0 Dec 27 08:10:03 crc kubenswrapper[4934]: I1227 08:10:03.815851 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e4cf02a3-796a-450f-a150-5ae72e0f6f7d","Type":"ContainerDied","Data":"c715a325324352bc15756616883cddc8ed263dbac45506fef4fec90a2dcf9bf2"} Dec 27 08:10:03 crc kubenswrapper[4934]: I1227 08:10:03.815879 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e4cf02a3-796a-450f-a150-5ae72e0f6f7d","Type":"ContainerDied","Data":"57c30729eec9072ee65d52f9e8e83f6cd9a51139aa6062e2adfb001dfc2bb00e"} Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.328412 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mrdpt" Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.451763 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/61f9c149-e7a9-4ca5-b708-aa9274ee74af-ssh-key\") pod \"61f9c149-e7a9-4ca5-b708-aa9274ee74af\" (UID: \"61f9c149-e7a9-4ca5-b708-aa9274ee74af\") " Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.451811 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f66kk\" (UniqueName: \"kubernetes.io/projected/61f9c149-e7a9-4ca5-b708-aa9274ee74af-kube-api-access-f66kk\") pod \"61f9c149-e7a9-4ca5-b708-aa9274ee74af\" (UID: \"61f9c149-e7a9-4ca5-b708-aa9274ee74af\") " Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.451891 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61f9c149-e7a9-4ca5-b708-aa9274ee74af-inventory\") pod \"61f9c149-e7a9-4ca5-b708-aa9274ee74af\" (UID: \"61f9c149-e7a9-4ca5-b708-aa9274ee74af\") " Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.458921 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61f9c149-e7a9-4ca5-b708-aa9274ee74af-kube-api-access-f66kk" (OuterVolumeSpecName: "kube-api-access-f66kk") pod "61f9c149-e7a9-4ca5-b708-aa9274ee74af" (UID: "61f9c149-e7a9-4ca5-b708-aa9274ee74af"). InnerVolumeSpecName "kube-api-access-f66kk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.491282 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61f9c149-e7a9-4ca5-b708-aa9274ee74af-inventory" (OuterVolumeSpecName: "inventory") pod "61f9c149-e7a9-4ca5-b708-aa9274ee74af" (UID: "61f9c149-e7a9-4ca5-b708-aa9274ee74af"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.503933 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61f9c149-e7a9-4ca5-b708-aa9274ee74af-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "61f9c149-e7a9-4ca5-b708-aa9274ee74af" (UID: "61f9c149-e7a9-4ca5-b708-aa9274ee74af"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.555839 4934 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/61f9c149-e7a9-4ca5-b708-aa9274ee74af-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.555869 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f66kk\" (UniqueName: \"kubernetes.io/projected/61f9c149-e7a9-4ca5-b708-aa9274ee74af-kube-api-access-f66kk\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.555881 4934 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/61f9c149-e7a9-4ca5-b708-aa9274ee74af-inventory\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.840013 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mrdpt" event={"ID":"61f9c149-e7a9-4ca5-b708-aa9274ee74af","Type":"ContainerDied","Data":"f1a7159e62e960762d4ae6c1ee01b5670c29da309af80b6dbd9e9b6f53a368e0"} Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.840345 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1a7159e62e960762d4ae6c1ee01b5670c29da309af80b6dbd9e9b6f53a368e0" Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.840200 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-mrdpt" Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.924201 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8"] Dec 27 08:10:05 crc kubenswrapper[4934]: E1227 08:10:05.924775 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f05689e-1351-435f-a267-f0c6508f1f4a" containerName="aodh-db-sync" Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.924797 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f05689e-1351-435f-a267-f0c6508f1f4a" containerName="aodh-db-sync" Dec 27 08:10:05 crc kubenswrapper[4934]: E1227 08:10:05.924851 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61f9c149-e7a9-4ca5-b708-aa9274ee74af" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.924861 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="61f9c149-e7a9-4ca5-b708-aa9274ee74af" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.925146 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f05689e-1351-435f-a267-f0c6508f1f4a" containerName="aodh-db-sync" Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.925181 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="61f9c149-e7a9-4ca5-b708-aa9274ee74af" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.926179 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8" Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.933219 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.933435 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.934522 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-98ddn" Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.944810 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8"] Dec 27 08:10:05 crc kubenswrapper[4934]: I1227 08:10:05.946610 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 27 08:10:06 crc kubenswrapper[4934]: I1227 08:10:06.068871 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8jf2\" (UniqueName: \"kubernetes.io/projected/68f2fdda-08a1-457f-b61e-c6d68ea69be3-kube-api-access-f8jf2\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8\" (UID: \"68f2fdda-08a1-457f-b61e-c6d68ea69be3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8" Dec 27 08:10:06 crc kubenswrapper[4934]: I1227 08:10:06.068967 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68f2fdda-08a1-457f-b61e-c6d68ea69be3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8\" (UID: \"68f2fdda-08a1-457f-b61e-c6d68ea69be3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8" Dec 27 08:10:06 crc kubenswrapper[4934]: I1227 08:10:06.069118 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68f2fdda-08a1-457f-b61e-c6d68ea69be3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8\" (UID: \"68f2fdda-08a1-457f-b61e-c6d68ea69be3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8" Dec 27 08:10:06 crc kubenswrapper[4934]: I1227 08:10:06.069178 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68f2fdda-08a1-457f-b61e-c6d68ea69be3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8\" (UID: \"68f2fdda-08a1-457f-b61e-c6d68ea69be3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8" Dec 27 08:10:06 crc kubenswrapper[4934]: I1227 08:10:06.171882 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68f2fdda-08a1-457f-b61e-c6d68ea69be3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8\" (UID: \"68f2fdda-08a1-457f-b61e-c6d68ea69be3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8" Dec 27 08:10:06 crc kubenswrapper[4934]: I1227 08:10:06.171972 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68f2fdda-08a1-457f-b61e-c6d68ea69be3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8\" (UID: \"68f2fdda-08a1-457f-b61e-c6d68ea69be3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8" Dec 27 08:10:06 crc kubenswrapper[4934]: I1227 08:10:06.172144 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8jf2\" (UniqueName: \"kubernetes.io/projected/68f2fdda-08a1-457f-b61e-c6d68ea69be3-kube-api-access-f8jf2\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8\" (UID: \"68f2fdda-08a1-457f-b61e-c6d68ea69be3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8" Dec 27 08:10:06 crc kubenswrapper[4934]: I1227 08:10:06.172199 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68f2fdda-08a1-457f-b61e-c6d68ea69be3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8\" (UID: \"68f2fdda-08a1-457f-b61e-c6d68ea69be3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8" Dec 27 08:10:06 crc kubenswrapper[4934]: I1227 08:10:06.177040 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68f2fdda-08a1-457f-b61e-c6d68ea69be3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8\" (UID: \"68f2fdda-08a1-457f-b61e-c6d68ea69be3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8" Dec 27 08:10:06 crc kubenswrapper[4934]: I1227 08:10:06.177609 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68f2fdda-08a1-457f-b61e-c6d68ea69be3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8\" (UID: \"68f2fdda-08a1-457f-b61e-c6d68ea69be3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8" Dec 27 08:10:06 crc kubenswrapper[4934]: I1227 08:10:06.177864 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68f2fdda-08a1-457f-b61e-c6d68ea69be3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8\" (UID: \"68f2fdda-08a1-457f-b61e-c6d68ea69be3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8" Dec 27 08:10:06 crc kubenswrapper[4934]: I1227 08:10:06.200865 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8jf2\" (UniqueName: \"kubernetes.io/projected/68f2fdda-08a1-457f-b61e-c6d68ea69be3-kube-api-access-f8jf2\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8\" (UID: \"68f2fdda-08a1-457f-b61e-c6d68ea69be3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8" Dec 27 08:10:06 crc kubenswrapper[4934]: I1227 08:10:06.289122 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8" Dec 27 08:10:06 crc kubenswrapper[4934]: I1227 08:10:06.881869 4934 generic.go:334] "Generic (PLEG): container finished" podID="e4cf02a3-796a-450f-a150-5ae72e0f6f7d" containerID="4e22b411703fbd8da59ac24c7542b39c88cde422839e82708b596306477ee2c0" exitCode=0 Dec 27 08:10:06 crc kubenswrapper[4934]: I1227 08:10:06.882150 4934 generic.go:334] "Generic (PLEG): container finished" podID="e4cf02a3-796a-450f-a150-5ae72e0f6f7d" containerID="95df1e1ab27861114324155498743dc1691daf87b8293aef64c5b5f124040970" exitCode=0 Dec 27 08:10:06 crc kubenswrapper[4934]: I1227 08:10:06.882172 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e4cf02a3-796a-450f-a150-5ae72e0f6f7d","Type":"ContainerDied","Data":"4e22b411703fbd8da59ac24c7542b39c88cde422839e82708b596306477ee2c0"} Dec 27 08:10:06 crc kubenswrapper[4934]: I1227 08:10:06.882198 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e4cf02a3-796a-450f-a150-5ae72e0f6f7d","Type":"ContainerDied","Data":"95df1e1ab27861114324155498743dc1691daf87b8293aef64c5b5f124040970"} Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.070932 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.094067 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8"] Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.198273 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-public-tls-certs\") pod \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.198393 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-config-data\") pod \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.198479 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-combined-ca-bundle\") pod \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.198539 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-scripts\") pod \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.198564 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-internal-tls-certs\") pod \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.198589 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5shk\" (UniqueName: \"kubernetes.io/projected/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-kube-api-access-g5shk\") pod \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\" (UID: \"e4cf02a3-796a-450f-a150-5ae72e0f6f7d\") " Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.204454 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-scripts" (OuterVolumeSpecName: "scripts") pod "e4cf02a3-796a-450f-a150-5ae72e0f6f7d" (UID: "e4cf02a3-796a-450f-a150-5ae72e0f6f7d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.205495 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-kube-api-access-g5shk" (OuterVolumeSpecName: "kube-api-access-g5shk") pod "e4cf02a3-796a-450f-a150-5ae72e0f6f7d" (UID: "e4cf02a3-796a-450f-a150-5ae72e0f6f7d"). InnerVolumeSpecName "kube-api-access-g5shk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.286569 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e4cf02a3-796a-450f-a150-5ae72e0f6f7d" (UID: "e4cf02a3-796a-450f-a150-5ae72e0f6f7d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.287184 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e4cf02a3-796a-450f-a150-5ae72e0f6f7d" (UID: "e4cf02a3-796a-450f-a150-5ae72e0f6f7d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.300949 4934 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.300979 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.300990 4934 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.300999 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5shk\" (UniqueName: \"kubernetes.io/projected/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-kube-api-access-g5shk\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.343230 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e4cf02a3-796a-450f-a150-5ae72e0f6f7d" (UID: "e4cf02a3-796a-450f-a150-5ae72e0f6f7d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.364514 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-config-data" (OuterVolumeSpecName: "config-data") pod "e4cf02a3-796a-450f-a150-5ae72e0f6f7d" (UID: "e4cf02a3-796a-450f-a150-5ae72e0f6f7d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.403484 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.403544 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4cf02a3-796a-450f-a150-5ae72e0f6f7d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.468023 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:10:07 crc kubenswrapper[4934]: E1227 08:10:07.468462 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.531892 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-1" podUID="5788f6c6-2500-4c66-af7a-830cde17caa4" containerName="rabbitmq" containerID="cri-o://e635ca483a8f18a75caaaa57d2ae4d90d80def0795952c3144ec97d276df4322" gracePeriod=604796 Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.894491 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8" event={"ID":"68f2fdda-08a1-457f-b61e-c6d68ea69be3","Type":"ContainerStarted","Data":"14d7cfeda23d75c493a93c5ae393ee825df59024358ca93161c9539286658d13"} Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.894552 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8" event={"ID":"68f2fdda-08a1-457f-b61e-c6d68ea69be3","Type":"ContainerStarted","Data":"51d7c6c5e1379aa66eb84e9cf98f47745c29e81607b2d537cede28de3c4b0c35"} Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.898938 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"e4cf02a3-796a-450f-a150-5ae72e0f6f7d","Type":"ContainerDied","Data":"8122585b52394ca06feed270d0e8e767cfe6799284964935b40bf60d6595cc13"} Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.898995 4934 scope.go:117] "RemoveContainer" containerID="4e22b411703fbd8da59ac24c7542b39c88cde422839e82708b596306477ee2c0" Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.899061 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.928440 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8" podStartSLOduration=2.521835734 podStartE2EDuration="2.928422842s" podCreationTimestamp="2025-12-27 08:10:05 +0000 UTC" firstStartedPulling="2025-12-27 08:10:07.131997372 +0000 UTC m=+1667.952437966" lastFinishedPulling="2025-12-27 08:10:07.53858448 +0000 UTC m=+1668.359025074" observedRunningTime="2025-12-27 08:10:07.909879092 +0000 UTC m=+1668.730319676" watchObservedRunningTime="2025-12-27 08:10:07.928422842 +0000 UTC m=+1668.748863436" Dec 27 08:10:07 crc kubenswrapper[4934]: I1227 08:10:07.987200 4934 scope.go:117] "RemoveContainer" containerID="95df1e1ab27861114324155498743dc1691daf87b8293aef64c5b5f124040970" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.024751 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.042148 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.054635 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 27 08:10:08 crc kubenswrapper[4934]: E1227 08:10:08.055220 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4cf02a3-796a-450f-a150-5ae72e0f6f7d" containerName="aodh-api" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.055239 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4cf02a3-796a-450f-a150-5ae72e0f6f7d" containerName="aodh-api" Dec 27 08:10:08 crc kubenswrapper[4934]: E1227 08:10:08.055278 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4cf02a3-796a-450f-a150-5ae72e0f6f7d" containerName="aodh-notifier" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.055284 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4cf02a3-796a-450f-a150-5ae72e0f6f7d" containerName="aodh-notifier" Dec 27 08:10:08 crc kubenswrapper[4934]: E1227 08:10:08.055308 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4cf02a3-796a-450f-a150-5ae72e0f6f7d" containerName="aodh-evaluator" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.055315 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4cf02a3-796a-450f-a150-5ae72e0f6f7d" containerName="aodh-evaluator" Dec 27 08:10:08 crc kubenswrapper[4934]: E1227 08:10:08.055338 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4cf02a3-796a-450f-a150-5ae72e0f6f7d" containerName="aodh-listener" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.055343 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4cf02a3-796a-450f-a150-5ae72e0f6f7d" containerName="aodh-listener" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.055550 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4cf02a3-796a-450f-a150-5ae72e0f6f7d" containerName="aodh-notifier" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.055579 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4cf02a3-796a-450f-a150-5ae72e0f6f7d" containerName="aodh-evaluator" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.055586 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4cf02a3-796a-450f-a150-5ae72e0f6f7d" containerName="aodh-api" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.055604 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4cf02a3-796a-450f-a150-5ae72e0f6f7d" containerName="aodh-listener" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.060448 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.064742 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-wrr58" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.064888 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.064904 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.065232 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.068473 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.070327 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.135002 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bj789\" (UniqueName: \"kubernetes.io/projected/ca41a5da-70d4-4668-ac75-978e192e46f4-kube-api-access-bj789\") pod \"aodh-0\" (UID: \"ca41a5da-70d4-4668-ac75-978e192e46f4\") " pod="openstack/aodh-0" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.135252 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca41a5da-70d4-4668-ac75-978e192e46f4-internal-tls-certs\") pod \"aodh-0\" (UID: \"ca41a5da-70d4-4668-ac75-978e192e46f4\") " pod="openstack/aodh-0" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.135289 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca41a5da-70d4-4668-ac75-978e192e46f4-combined-ca-bundle\") pod \"aodh-0\" (UID: \"ca41a5da-70d4-4668-ac75-978e192e46f4\") " pod="openstack/aodh-0" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.135534 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca41a5da-70d4-4668-ac75-978e192e46f4-public-tls-certs\") pod \"aodh-0\" (UID: \"ca41a5da-70d4-4668-ac75-978e192e46f4\") " pod="openstack/aodh-0" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.135596 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca41a5da-70d4-4668-ac75-978e192e46f4-config-data\") pod \"aodh-0\" (UID: \"ca41a5da-70d4-4668-ac75-978e192e46f4\") " pod="openstack/aodh-0" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.135623 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca41a5da-70d4-4668-ac75-978e192e46f4-scripts\") pod \"aodh-0\" (UID: \"ca41a5da-70d4-4668-ac75-978e192e46f4\") " pod="openstack/aodh-0" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.142750 4934 scope.go:117] "RemoveContainer" containerID="c715a325324352bc15756616883cddc8ed263dbac45506fef4fec90a2dcf9bf2" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.173328 4934 scope.go:117] "RemoveContainer" containerID="57c30729eec9072ee65d52f9e8e83f6cd9a51139aa6062e2adfb001dfc2bb00e" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.240958 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca41a5da-70d4-4668-ac75-978e192e46f4-public-tls-certs\") pod \"aodh-0\" (UID: \"ca41a5da-70d4-4668-ac75-978e192e46f4\") " pod="openstack/aodh-0" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.241044 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca41a5da-70d4-4668-ac75-978e192e46f4-config-data\") pod \"aodh-0\" (UID: \"ca41a5da-70d4-4668-ac75-978e192e46f4\") " pod="openstack/aodh-0" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.241112 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca41a5da-70d4-4668-ac75-978e192e46f4-scripts\") pod \"aodh-0\" (UID: \"ca41a5da-70d4-4668-ac75-978e192e46f4\") " pod="openstack/aodh-0" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.241226 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bj789\" (UniqueName: \"kubernetes.io/projected/ca41a5da-70d4-4668-ac75-978e192e46f4-kube-api-access-bj789\") pod \"aodh-0\" (UID: \"ca41a5da-70d4-4668-ac75-978e192e46f4\") " pod="openstack/aodh-0" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.241256 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca41a5da-70d4-4668-ac75-978e192e46f4-internal-tls-certs\") pod \"aodh-0\" (UID: \"ca41a5da-70d4-4668-ac75-978e192e46f4\") " pod="openstack/aodh-0" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.241307 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca41a5da-70d4-4668-ac75-978e192e46f4-combined-ca-bundle\") pod \"aodh-0\" (UID: \"ca41a5da-70d4-4668-ac75-978e192e46f4\") " pod="openstack/aodh-0" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.245882 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca41a5da-70d4-4668-ac75-978e192e46f4-internal-tls-certs\") pod \"aodh-0\" (UID: \"ca41a5da-70d4-4668-ac75-978e192e46f4\") " pod="openstack/aodh-0" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.246988 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca41a5da-70d4-4668-ac75-978e192e46f4-combined-ca-bundle\") pod \"aodh-0\" (UID: \"ca41a5da-70d4-4668-ac75-978e192e46f4\") " pod="openstack/aodh-0" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.247212 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca41a5da-70d4-4668-ac75-978e192e46f4-public-tls-certs\") pod \"aodh-0\" (UID: \"ca41a5da-70d4-4668-ac75-978e192e46f4\") " pod="openstack/aodh-0" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.247461 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca41a5da-70d4-4668-ac75-978e192e46f4-config-data\") pod \"aodh-0\" (UID: \"ca41a5da-70d4-4668-ac75-978e192e46f4\") " pod="openstack/aodh-0" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.250406 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca41a5da-70d4-4668-ac75-978e192e46f4-scripts\") pod \"aodh-0\" (UID: \"ca41a5da-70d4-4668-ac75-978e192e46f4\") " pod="openstack/aodh-0" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.260053 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bj789\" (UniqueName: \"kubernetes.io/projected/ca41a5da-70d4-4668-ac75-978e192e46f4-kube-api-access-bj789\") pod \"aodh-0\" (UID: \"ca41a5da-70d4-4668-ac75-978e192e46f4\") " pod="openstack/aodh-0" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.383616 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.831679 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-1" podUID="5788f6c6-2500-4c66-af7a-830cde17caa4" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.130:5671: connect: connection refused" Dec 27 08:10:08 crc kubenswrapper[4934]: I1227 08:10:08.945366 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 27 08:10:09 crc kubenswrapper[4934]: I1227 08:10:09.484287 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4cf02a3-796a-450f-a150-5ae72e0f6f7d" path="/var/lib/kubelet/pods/e4cf02a3-796a-450f-a150-5ae72e0f6f7d/volumes" Dec 27 08:10:09 crc kubenswrapper[4934]: I1227 08:10:09.925184 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ca41a5da-70d4-4668-ac75-978e192e46f4","Type":"ContainerStarted","Data":"81a426937e8f567f75c92bedb2067e32c66870a15fda0245b4e7364dbb4f9013"} Dec 27 08:10:09 crc kubenswrapper[4934]: I1227 08:10:09.925495 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ca41a5da-70d4-4668-ac75-978e192e46f4","Type":"ContainerStarted","Data":"6549f0f3eae20e9c5d24440e3203585cf8faa848b407dfa3784e419001dc2f34"} Dec 27 08:10:10 crc kubenswrapper[4934]: I1227 08:10:10.941891 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ca41a5da-70d4-4668-ac75-978e192e46f4","Type":"ContainerStarted","Data":"ce21c018d9ad9faaef803cc71c0e34b021f05f7a8d2eb034143fc7a7d757a073"} Dec 27 08:10:12 crc kubenswrapper[4934]: I1227 08:10:12.974372 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ca41a5da-70d4-4668-ac75-978e192e46f4","Type":"ContainerStarted","Data":"6f73b98bad13ff1b2151087e5365f405fbbd81404a46dd80947958e54a6f464c"} Dec 27 08:10:13 crc kubenswrapper[4934]: I1227 08:10:13.993053 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"ca41a5da-70d4-4668-ac75-978e192e46f4","Type":"ContainerStarted","Data":"9ac2c71e9c06c3826af2bb40bf0738d9286120e7370ac67036683f69677f557c"} Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.002794 4934 generic.go:334] "Generic (PLEG): container finished" podID="5788f6c6-2500-4c66-af7a-830cde17caa4" containerID="e635ca483a8f18a75caaaa57d2ae4d90d80def0795952c3144ec97d276df4322" exitCode=0 Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.002831 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"5788f6c6-2500-4c66-af7a-830cde17caa4","Type":"ContainerDied","Data":"e635ca483a8f18a75caaaa57d2ae4d90d80def0795952c3144ec97d276df4322"} Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.015303 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.821817416 podStartE2EDuration="7.015285466s" podCreationTimestamp="2025-12-27 08:10:07 +0000 UTC" firstStartedPulling="2025-12-27 08:10:08.953570155 +0000 UTC m=+1669.774010769" lastFinishedPulling="2025-12-27 08:10:13.147038225 +0000 UTC m=+1673.967478819" observedRunningTime="2025-12-27 08:10:14.012327163 +0000 UTC m=+1674.832767797" watchObservedRunningTime="2025-12-27 08:10:14.015285466 +0000 UTC m=+1674.835726070" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.309635 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.395662 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8496x\" (UniqueName: \"kubernetes.io/projected/5788f6c6-2500-4c66-af7a-830cde17caa4-kube-api-access-8496x\") pod \"5788f6c6-2500-4c66-af7a-830cde17caa4\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.395757 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5788f6c6-2500-4c66-af7a-830cde17caa4-config-data\") pod \"5788f6c6-2500-4c66-af7a-830cde17caa4\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.395822 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-erlang-cookie\") pod \"5788f6c6-2500-4c66-af7a-830cde17caa4\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.395883 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-plugins\") pod \"5788f6c6-2500-4c66-af7a-830cde17caa4\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.395958 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-tls\") pod \"5788f6c6-2500-4c66-af7a-830cde17caa4\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.396574 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "5788f6c6-2500-4c66-af7a-830cde17caa4" (UID: "5788f6c6-2500-4c66-af7a-830cde17caa4"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.396968 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "5788f6c6-2500-4c66-af7a-830cde17caa4" (UID: "5788f6c6-2500-4c66-af7a-830cde17caa4"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.397035 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2\") pod \"5788f6c6-2500-4c66-af7a-830cde17caa4\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.397167 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-confd\") pod \"5788f6c6-2500-4c66-af7a-830cde17caa4\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.397258 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5788f6c6-2500-4c66-af7a-830cde17caa4-plugins-conf\") pod \"5788f6c6-2500-4c66-af7a-830cde17caa4\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.397307 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5788f6c6-2500-4c66-af7a-830cde17caa4-erlang-cookie-secret\") pod \"5788f6c6-2500-4c66-af7a-830cde17caa4\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.397349 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5788f6c6-2500-4c66-af7a-830cde17caa4-pod-info\") pod \"5788f6c6-2500-4c66-af7a-830cde17caa4\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.397445 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5788f6c6-2500-4c66-af7a-830cde17caa4-server-conf\") pod \"5788f6c6-2500-4c66-af7a-830cde17caa4\" (UID: \"5788f6c6-2500-4c66-af7a-830cde17caa4\") " Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.398469 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.398501 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.402700 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5788f6c6-2500-4c66-af7a-830cde17caa4-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "5788f6c6-2500-4c66-af7a-830cde17caa4" (UID: "5788f6c6-2500-4c66-af7a-830cde17caa4"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.404788 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "5788f6c6-2500-4c66-af7a-830cde17caa4" (UID: "5788f6c6-2500-4c66-af7a-830cde17caa4"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.407503 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5788f6c6-2500-4c66-af7a-830cde17caa4-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "5788f6c6-2500-4c66-af7a-830cde17caa4" (UID: "5788f6c6-2500-4c66-af7a-830cde17caa4"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.414022 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/5788f6c6-2500-4c66-af7a-830cde17caa4-pod-info" (OuterVolumeSpecName: "pod-info") pod "5788f6c6-2500-4c66-af7a-830cde17caa4" (UID: "5788f6c6-2500-4c66-af7a-830cde17caa4"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.417326 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5788f6c6-2500-4c66-af7a-830cde17caa4-kube-api-access-8496x" (OuterVolumeSpecName: "kube-api-access-8496x") pod "5788f6c6-2500-4c66-af7a-830cde17caa4" (UID: "5788f6c6-2500-4c66-af7a-830cde17caa4"). InnerVolumeSpecName "kube-api-access-8496x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.434678 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5788f6c6-2500-4c66-af7a-830cde17caa4-config-data" (OuterVolumeSpecName: "config-data") pod "5788f6c6-2500-4c66-af7a-830cde17caa4" (UID: "5788f6c6-2500-4c66-af7a-830cde17caa4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.448237 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2" (OuterVolumeSpecName: "persistence") pod "5788f6c6-2500-4c66-af7a-830cde17caa4" (UID: "5788f6c6-2500-4c66-af7a-830cde17caa4"). InnerVolumeSpecName "pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.469449 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5788f6c6-2500-4c66-af7a-830cde17caa4-server-conf" (OuterVolumeSpecName: "server-conf") pod "5788f6c6-2500-4c66-af7a-830cde17caa4" (UID: "5788f6c6-2500-4c66-af7a-830cde17caa4"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.500127 4934 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5788f6c6-2500-4c66-af7a-830cde17caa4-server-conf\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.500155 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8496x\" (UniqueName: \"kubernetes.io/projected/5788f6c6-2500-4c66-af7a-830cde17caa4-kube-api-access-8496x\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.500166 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5788f6c6-2500-4c66-af7a-830cde17caa4-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.500174 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.500194 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2\") on node \"crc\" " Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.500204 4934 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5788f6c6-2500-4c66-af7a-830cde17caa4-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.500216 4934 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5788f6c6-2500-4c66-af7a-830cde17caa4-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.500224 4934 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5788f6c6-2500-4c66-af7a-830cde17caa4-pod-info\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.529889 4934 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.530044 4934 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2") on node "crc" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.545260 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "5788f6c6-2500-4c66-af7a-830cde17caa4" (UID: "5788f6c6-2500-4c66-af7a-830cde17caa4"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.602696 4934 reconciler_common.go:293] "Volume detached for volume \"pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:14 crc kubenswrapper[4934]: I1227 08:10:14.602740 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5788f6c6-2500-4c66-af7a-830cde17caa4-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.016441 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"5788f6c6-2500-4c66-af7a-830cde17caa4","Type":"ContainerDied","Data":"f4003fb78d58c40b4d21809dc8d97be99a25f219927cd511063ec4feb9101b6f"} Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.016476 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.016513 4934 scope.go:117] "RemoveContainer" containerID="e635ca483a8f18a75caaaa57d2ae4d90d80def0795952c3144ec97d276df4322" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.043231 4934 scope.go:117] "RemoveContainer" containerID="025c6e94fc16f8dfe3303fd225b5fc2e02f9952414fc805476597fdb089f791b" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.070396 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-1"] Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.096098 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-1"] Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.107539 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-1"] Dec 27 08:10:15 crc kubenswrapper[4934]: E1227 08:10:15.108696 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5788f6c6-2500-4c66-af7a-830cde17caa4" containerName="rabbitmq" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.108720 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="5788f6c6-2500-4c66-af7a-830cde17caa4" containerName="rabbitmq" Dec 27 08:10:15 crc kubenswrapper[4934]: E1227 08:10:15.108734 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5788f6c6-2500-4c66-af7a-830cde17caa4" containerName="setup-container" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.108744 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="5788f6c6-2500-4c66-af7a-830cde17caa4" containerName="setup-container" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.108993 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="5788f6c6-2500-4c66-af7a-830cde17caa4" containerName="rabbitmq" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.110409 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.157632 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-1"] Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.215212 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93d5ae97-eb77-4368-931c-0421b65ac057-config-data\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.215301 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/93d5ae97-eb77-4368-931c-0421b65ac057-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.215348 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/93d5ae97-eb77-4368-931c-0421b65ac057-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.215388 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-624qg\" (UniqueName: \"kubernetes.io/projected/93d5ae97-eb77-4368-931c-0421b65ac057-kube-api-access-624qg\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.215410 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/93d5ae97-eb77-4368-931c-0421b65ac057-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.215455 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/93d5ae97-eb77-4368-931c-0421b65ac057-pod-info\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.215474 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/93d5ae97-eb77-4368-931c-0421b65ac057-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.215488 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/93d5ae97-eb77-4368-931c-0421b65ac057-server-conf\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.215504 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.215541 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/93d5ae97-eb77-4368-931c-0421b65ac057-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.215559 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/93d5ae97-eb77-4368-931c-0421b65ac057-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.317796 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/93d5ae97-eb77-4368-931c-0421b65ac057-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.317881 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/93d5ae97-eb77-4368-931c-0421b65ac057-pod-info\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.317904 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/93d5ae97-eb77-4368-931c-0421b65ac057-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.317922 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/93d5ae97-eb77-4368-931c-0421b65ac057-server-conf\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.317941 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.317976 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/93d5ae97-eb77-4368-931c-0421b65ac057-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.317993 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/93d5ae97-eb77-4368-931c-0421b65ac057-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.318036 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93d5ae97-eb77-4368-931c-0421b65ac057-config-data\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.318104 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/93d5ae97-eb77-4368-931c-0421b65ac057-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.318146 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/93d5ae97-eb77-4368-931c-0421b65ac057-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.318186 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-624qg\" (UniqueName: \"kubernetes.io/projected/93d5ae97-eb77-4368-931c-0421b65ac057-kube-api-access-624qg\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.320327 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/93d5ae97-eb77-4368-931c-0421b65ac057-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.320394 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/93d5ae97-eb77-4368-931c-0421b65ac057-server-conf\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.320631 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/93d5ae97-eb77-4368-931c-0421b65ac057-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.320990 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93d5ae97-eb77-4368-931c-0421b65ac057-config-data\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.321378 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/93d5ae97-eb77-4368-931c-0421b65ac057-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.324713 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/93d5ae97-eb77-4368-931c-0421b65ac057-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.324959 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/93d5ae97-eb77-4368-931c-0421b65ac057-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.326144 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/93d5ae97-eb77-4368-931c-0421b65ac057-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.327653 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/93d5ae97-eb77-4368-931c-0421b65ac057-pod-info\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.339776 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-624qg\" (UniqueName: \"kubernetes.io/projected/93d5ae97-eb77-4368-931c-0421b65ac057-kube-api-access-624qg\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.479604 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5788f6c6-2500-4c66-af7a-830cde17caa4" path="/var/lib/kubelet/pods/5788f6c6-2500-4c66-af7a-830cde17caa4/volumes" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.497057 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.497122 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/faea16fb88df6ea32d8bc003326a02149652e41cb41067608e5657db52571d30/globalmount\"" pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.552064 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-60a63cfe-8823-4b62-95a2-7478bd7d38a2\") pod \"rabbitmq-server-1\" (UID: \"93d5ae97-eb77-4368-931c-0421b65ac057\") " pod="openstack/rabbitmq-server-1" Dec 27 08:10:15 crc kubenswrapper[4934]: I1227 08:10:15.558828 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Dec 27 08:10:16 crc kubenswrapper[4934]: I1227 08:10:16.052900 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-1"] Dec 27 08:10:16 crc kubenswrapper[4934]: W1227 08:10:16.055199 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93d5ae97_eb77_4368_931c_0421b65ac057.slice/crio-386306e5100ad1b9e1a7a58959cf6b1bd5b2c78de8bd0cf3a27cf6b5e7fcb89c WatchSource:0}: Error finding container 386306e5100ad1b9e1a7a58959cf6b1bd5b2c78de8bd0cf3a27cf6b5e7fcb89c: Status 404 returned error can't find the container with id 386306e5100ad1b9e1a7a58959cf6b1bd5b2c78de8bd0cf3a27cf6b5e7fcb89c Dec 27 08:10:17 crc kubenswrapper[4934]: I1227 08:10:17.044975 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"93d5ae97-eb77-4368-931c-0421b65ac057","Type":"ContainerStarted","Data":"386306e5100ad1b9e1a7a58959cf6b1bd5b2c78de8bd0cf3a27cf6b5e7fcb89c"} Dec 27 08:10:19 crc kubenswrapper[4934]: I1227 08:10:19.111747 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"93d5ae97-eb77-4368-931c-0421b65ac057","Type":"ContainerStarted","Data":"2982c3ffd77f705b05b2879fe17427d629579a427107202ee3a193dddb77c4a1"} Dec 27 08:10:22 crc kubenswrapper[4934]: I1227 08:10:22.468404 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:10:22 crc kubenswrapper[4934]: E1227 08:10:22.469004 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:10:24 crc kubenswrapper[4934]: I1227 08:10:24.293976 4934 scope.go:117] "RemoveContainer" containerID="032cc47dedce6dbbea921dd0ef1607d653ed430218ac0956abc28d42c32b092e" Dec 27 08:10:24 crc kubenswrapper[4934]: I1227 08:10:24.350730 4934 scope.go:117] "RemoveContainer" containerID="f7124ee43fecbe067924a0a5202f619026d658a023e3c1209dc7581a192ef658" Dec 27 08:10:24 crc kubenswrapper[4934]: I1227 08:10:24.429723 4934 scope.go:117] "RemoveContainer" containerID="5805e834233bfbef53115169349dbc519f2b820afe8d0bea2e11c01c167f724c" Dec 27 08:10:24 crc kubenswrapper[4934]: I1227 08:10:24.482395 4934 scope.go:117] "RemoveContainer" containerID="aac373063ba829ef3ed599024d2a013dd99da808d9e914e732e6fc9b32afa8fd" Dec 27 08:10:37 crc kubenswrapper[4934]: I1227 08:10:37.467552 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:10:37 crc kubenswrapper[4934]: E1227 08:10:37.468577 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:10:50 crc kubenswrapper[4934]: I1227 08:10:50.469432 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:10:50 crc kubenswrapper[4934]: E1227 08:10:50.470988 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:10:51 crc kubenswrapper[4934]: I1227 08:10:51.678285 4934 generic.go:334] "Generic (PLEG): container finished" podID="93d5ae97-eb77-4368-931c-0421b65ac057" containerID="2982c3ffd77f705b05b2879fe17427d629579a427107202ee3a193dddb77c4a1" exitCode=0 Dec 27 08:10:51 crc kubenswrapper[4934]: I1227 08:10:51.678387 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"93d5ae97-eb77-4368-931c-0421b65ac057","Type":"ContainerDied","Data":"2982c3ffd77f705b05b2879fe17427d629579a427107202ee3a193dddb77c4a1"} Dec 27 08:10:52 crc kubenswrapper[4934]: I1227 08:10:52.694743 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"93d5ae97-eb77-4368-931c-0421b65ac057","Type":"ContainerStarted","Data":"46aa620518a7d04acfeac31db56afc1860ae4189e6c12a630ab495b34a7b739d"} Dec 27 08:10:52 crc kubenswrapper[4934]: I1227 08:10:52.696223 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-1" Dec 27 08:10:52 crc kubenswrapper[4934]: I1227 08:10:52.737545 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-1" podStartSLOduration=37.737526383 podStartE2EDuration="37.737526383s" podCreationTimestamp="2025-12-27 08:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:10:52.725354061 +0000 UTC m=+1713.545794665" watchObservedRunningTime="2025-12-27 08:10:52.737526383 +0000 UTC m=+1713.557966977" Dec 27 08:11:03 crc kubenswrapper[4934]: I1227 08:11:03.467700 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:11:03 crc kubenswrapper[4934]: E1227 08:11:03.468903 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:11:05 crc kubenswrapper[4934]: I1227 08:11:05.562400 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-1" Dec 27 08:11:05 crc kubenswrapper[4934]: I1227 08:11:05.677927 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 27 08:11:09 crc kubenswrapper[4934]: I1227 08:11:09.788445 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c" containerName="rabbitmq" containerID="cri-o://00e3619755c94f9d80fd73245a68300e5f89f6b0603f9ba2025aa1302d112f7d" gracePeriod=604796 Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.593224 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.662390 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-confd\") pod \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.662535 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-server-conf\") pod \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.662939 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-919d5745-3220-439b-a74b-98d67847c8df\") pod \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.663005 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-plugins\") pod \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.663036 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6s5mg\" (UniqueName: \"kubernetes.io/projected/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-kube-api-access-6s5mg\") pod \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.663056 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-erlang-cookie\") pod \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.663157 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-config-data\") pod \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.663190 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-tls\") pod \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.663246 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-erlang-cookie-secret\") pod \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.663283 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-plugins-conf\") pod \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.663324 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-pod-info\") pod \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\" (UID: \"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c\") " Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.672449 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-kube-api-access-6s5mg" (OuterVolumeSpecName: "kube-api-access-6s5mg") pod "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c" (UID: "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c"). InnerVolumeSpecName "kube-api-access-6s5mg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.673148 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c" (UID: "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.673346 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-pod-info" (OuterVolumeSpecName: "pod-info") pod "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c" (UID: "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.674148 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c" (UID: "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.675911 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c" (UID: "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.676325 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c" (UID: "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.681987 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c" (UID: "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.755570 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-config-data" (OuterVolumeSpecName: "config-data") pod "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c" (UID: "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.767154 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6s5mg\" (UniqueName: \"kubernetes.io/projected/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-kube-api-access-6s5mg\") on node \"crc\" DevicePath \"\"" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.767201 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.767213 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.767225 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.767237 4934 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.767248 4934 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.767261 4934 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-pod-info\") on node \"crc\" DevicePath \"\"" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.767271 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.767149 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-919d5745-3220-439b-a74b-98d67847c8df" (OuterVolumeSpecName: "persistence") pod "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c" (UID: "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c"). InnerVolumeSpecName "pvc-919d5745-3220-439b-a74b-98d67847c8df". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.784421 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-server-conf" (OuterVolumeSpecName: "server-conf") pod "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c" (UID: "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.854983 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c" (UID: "cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.868969 4934 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.868999 4934 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c-server-conf\") on node \"crc\" DevicePath \"\"" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.869034 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-919d5745-3220-439b-a74b-98d67847c8df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-919d5745-3220-439b-a74b-98d67847c8df\") on node \"crc\" " Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.914737 4934 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.914923 4934 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-919d5745-3220-439b-a74b-98d67847c8df" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-919d5745-3220-439b-a74b-98d67847c8df") on node "crc" Dec 27 08:11:16 crc kubenswrapper[4934]: I1227 08:11:16.971430 4934 reconciler_common.go:293] "Volume detached for volume \"pvc-919d5745-3220-439b-a74b-98d67847c8df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-919d5745-3220-439b-a74b-98d67847c8df\") on node \"crc\" DevicePath \"\"" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.021053 4934 generic.go:334] "Generic (PLEG): container finished" podID="cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c" containerID="00e3619755c94f9d80fd73245a68300e5f89f6b0603f9ba2025aa1302d112f7d" exitCode=0 Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.021121 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.021128 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c","Type":"ContainerDied","Data":"00e3619755c94f9d80fd73245a68300e5f89f6b0603f9ba2025aa1302d112f7d"} Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.021170 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c","Type":"ContainerDied","Data":"3f2372d8ca89e99aac8a9574dbe4a130bc8c572cfa304e120ad6e5f99f44cc42"} Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.021188 4934 scope.go:117] "RemoveContainer" containerID="00e3619755c94f9d80fd73245a68300e5f89f6b0603f9ba2025aa1302d112f7d" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.068222 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.074341 4934 scope.go:117] "RemoveContainer" containerID="9e896bff961270fe6bc84ed6b42b91ba71952e4040cd6b4594ed7c4e6be8c873" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.085137 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.101468 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 27 08:11:17 crc kubenswrapper[4934]: E1227 08:11:17.102014 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c" containerName="rabbitmq" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.102027 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c" containerName="rabbitmq" Dec 27 08:11:17 crc kubenswrapper[4934]: E1227 08:11:17.102049 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c" containerName="setup-container" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.102055 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c" containerName="setup-container" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.102300 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c" containerName="rabbitmq" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.103635 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.132157 4934 scope.go:117] "RemoveContainer" containerID="00e3619755c94f9d80fd73245a68300e5f89f6b0603f9ba2025aa1302d112f7d" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.132380 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 27 08:11:17 crc kubenswrapper[4934]: E1227 08:11:17.136520 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00e3619755c94f9d80fd73245a68300e5f89f6b0603f9ba2025aa1302d112f7d\": container with ID starting with 00e3619755c94f9d80fd73245a68300e5f89f6b0603f9ba2025aa1302d112f7d not found: ID does not exist" containerID="00e3619755c94f9d80fd73245a68300e5f89f6b0603f9ba2025aa1302d112f7d" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.136580 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00e3619755c94f9d80fd73245a68300e5f89f6b0603f9ba2025aa1302d112f7d"} err="failed to get container status \"00e3619755c94f9d80fd73245a68300e5f89f6b0603f9ba2025aa1302d112f7d\": rpc error: code = NotFound desc = could not find container \"00e3619755c94f9d80fd73245a68300e5f89f6b0603f9ba2025aa1302d112f7d\": container with ID starting with 00e3619755c94f9d80fd73245a68300e5f89f6b0603f9ba2025aa1302d112f7d not found: ID does not exist" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.136617 4934 scope.go:117] "RemoveContainer" containerID="9e896bff961270fe6bc84ed6b42b91ba71952e4040cd6b4594ed7c4e6be8c873" Dec 27 08:11:17 crc kubenswrapper[4934]: E1227 08:11:17.140437 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e896bff961270fe6bc84ed6b42b91ba71952e4040cd6b4594ed7c4e6be8c873\": container with ID starting with 9e896bff961270fe6bc84ed6b42b91ba71952e4040cd6b4594ed7c4e6be8c873 not found: ID does not exist" containerID="9e896bff961270fe6bc84ed6b42b91ba71952e4040cd6b4594ed7c4e6be8c873" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.140498 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e896bff961270fe6bc84ed6b42b91ba71952e4040cd6b4594ed7c4e6be8c873"} err="failed to get container status \"9e896bff961270fe6bc84ed6b42b91ba71952e4040cd6b4594ed7c4e6be8c873\": rpc error: code = NotFound desc = could not find container \"9e896bff961270fe6bc84ed6b42b91ba71952e4040cd6b4594ed7c4e6be8c873\": container with ID starting with 9e896bff961270fe6bc84ed6b42b91ba71952e4040cd6b4594ed7c4e6be8c873 not found: ID does not exist" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.176875 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-919d5745-3220-439b-a74b-98d67847c8df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-919d5745-3220-439b-a74b-98d67847c8df\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.176941 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.176974 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.176994 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.177013 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.177036 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.177282 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-config-data\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.177319 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.177346 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfmrx\" (UniqueName: \"kubernetes.io/projected/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-kube-api-access-lfmrx\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.177527 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.177775 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.282580 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.282717 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.282788 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-919d5745-3220-439b-a74b-98d67847c8df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-919d5745-3220-439b-a74b-98d67847c8df\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.282855 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.282890 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.282921 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.282949 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.282981 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.283135 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-config-data\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.283171 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.283198 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfmrx\" (UniqueName: \"kubernetes.io/projected/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-kube-api-access-lfmrx\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.286811 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.287974 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.296370 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.296617 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.302122 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.310371 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfmrx\" (UniqueName: \"kubernetes.io/projected/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-kube-api-access-lfmrx\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.313864 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.314015 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-config-data\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.315506 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.317920 4934 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.317985 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-919d5745-3220-439b-a74b-98d67847c8df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-919d5745-3220-439b-a74b-98d67847c8df\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d634143c1674768228e9390fde62df5502f89e4584183c04f80432b07343da17/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.321300 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b6a52bce-47cf-48e4-9f9a-7b3d978dee58-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.382883 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-919d5745-3220-439b-a74b-98d67847c8df\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-919d5745-3220-439b-a74b-98d67847c8df\") pod \"rabbitmq-server-0\" (UID: \"b6a52bce-47cf-48e4-9f9a-7b3d978dee58\") " pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.492813 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 27 08:11:17 crc kubenswrapper[4934]: I1227 08:11:17.498544 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c" path="/var/lib/kubelet/pods/cd9b61ad-7e6b-4ddd-b31b-b32ca0ad9a4c/volumes" Dec 27 08:11:18 crc kubenswrapper[4934]: I1227 08:11:18.028770 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 27 08:11:18 crc kubenswrapper[4934]: I1227 08:11:18.468124 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:11:18 crc kubenswrapper[4934]: E1227 08:11:18.469321 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:11:19 crc kubenswrapper[4934]: I1227 08:11:19.050910 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b6a52bce-47cf-48e4-9f9a-7b3d978dee58","Type":"ContainerStarted","Data":"2245d5c3deec0d14fb3e4c6ac1f0a5bee519cbff4a15dc249afed578a9cbcf3d"} Dec 27 08:11:21 crc kubenswrapper[4934]: I1227 08:11:21.096027 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b6a52bce-47cf-48e4-9f9a-7b3d978dee58","Type":"ContainerStarted","Data":"eacb688b9fb55bcf4574776a4cea69f298a2fb0b5e0a4624536eef68f53e492b"} Dec 27 08:11:24 crc kubenswrapper[4934]: I1227 08:11:24.805893 4934 scope.go:117] "RemoveContainer" containerID="db4ed2f28d2e711d823b1ea76d20bc93ee4ecdf738544c01eaea7df7995d7b4a" Dec 27 08:11:30 crc kubenswrapper[4934]: I1227 08:11:30.468533 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:11:30 crc kubenswrapper[4934]: E1227 08:11:30.469997 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:11:45 crc kubenswrapper[4934]: I1227 08:11:45.468337 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:11:45 crc kubenswrapper[4934]: E1227 08:11:45.469492 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:11:53 crc kubenswrapper[4934]: I1227 08:11:53.654690 4934 generic.go:334] "Generic (PLEG): container finished" podID="b6a52bce-47cf-48e4-9f9a-7b3d978dee58" containerID="eacb688b9fb55bcf4574776a4cea69f298a2fb0b5e0a4624536eef68f53e492b" exitCode=0 Dec 27 08:11:53 crc kubenswrapper[4934]: I1227 08:11:53.654829 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b6a52bce-47cf-48e4-9f9a-7b3d978dee58","Type":"ContainerDied","Data":"eacb688b9fb55bcf4574776a4cea69f298a2fb0b5e0a4624536eef68f53e492b"} Dec 27 08:11:54 crc kubenswrapper[4934]: I1227 08:11:54.677700 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b6a52bce-47cf-48e4-9f9a-7b3d978dee58","Type":"ContainerStarted","Data":"aeddb662be23c4d8162c4260742533ad02ec60bb31b5e4e1bfe6602d7a326713"} Dec 27 08:11:54 crc kubenswrapper[4934]: I1227 08:11:54.678772 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 27 08:11:54 crc kubenswrapper[4934]: I1227 08:11:54.724355 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.724328151 podStartE2EDuration="37.724328151s" podCreationTimestamp="2025-12-27 08:11:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:11:54.70892644 +0000 UTC m=+1775.529367074" watchObservedRunningTime="2025-12-27 08:11:54.724328151 +0000 UTC m=+1775.544768775" Dec 27 08:11:59 crc kubenswrapper[4934]: I1227 08:11:59.478693 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:11:59 crc kubenswrapper[4934]: E1227 08:11:59.479992 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:12:01 crc kubenswrapper[4934]: I1227 08:12:01.595024 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9zwp8"] Dec 27 08:12:01 crc kubenswrapper[4934]: I1227 08:12:01.599810 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9zwp8" Dec 27 08:12:01 crc kubenswrapper[4934]: I1227 08:12:01.608764 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9zwp8"] Dec 27 08:12:01 crc kubenswrapper[4934]: I1227 08:12:01.756978 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjzcb\" (UniqueName: \"kubernetes.io/projected/ff664db3-160a-4207-a543-9f43fff5e93a-kube-api-access-cjzcb\") pod \"community-operators-9zwp8\" (UID: \"ff664db3-160a-4207-a543-9f43fff5e93a\") " pod="openshift-marketplace/community-operators-9zwp8" Dec 27 08:12:01 crc kubenswrapper[4934]: I1227 08:12:01.757167 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff664db3-160a-4207-a543-9f43fff5e93a-utilities\") pod \"community-operators-9zwp8\" (UID: \"ff664db3-160a-4207-a543-9f43fff5e93a\") " pod="openshift-marketplace/community-operators-9zwp8" Dec 27 08:12:01 crc kubenswrapper[4934]: I1227 08:12:01.757235 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff664db3-160a-4207-a543-9f43fff5e93a-catalog-content\") pod \"community-operators-9zwp8\" (UID: \"ff664db3-160a-4207-a543-9f43fff5e93a\") " pod="openshift-marketplace/community-operators-9zwp8" Dec 27 08:12:01 crc kubenswrapper[4934]: I1227 08:12:01.860679 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjzcb\" (UniqueName: \"kubernetes.io/projected/ff664db3-160a-4207-a543-9f43fff5e93a-kube-api-access-cjzcb\") pod \"community-operators-9zwp8\" (UID: \"ff664db3-160a-4207-a543-9f43fff5e93a\") " pod="openshift-marketplace/community-operators-9zwp8" Dec 27 08:12:01 crc kubenswrapper[4934]: I1227 08:12:01.860954 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff664db3-160a-4207-a543-9f43fff5e93a-utilities\") pod \"community-operators-9zwp8\" (UID: \"ff664db3-160a-4207-a543-9f43fff5e93a\") " pod="openshift-marketplace/community-operators-9zwp8" Dec 27 08:12:01 crc kubenswrapper[4934]: I1227 08:12:01.861072 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff664db3-160a-4207-a543-9f43fff5e93a-catalog-content\") pod \"community-operators-9zwp8\" (UID: \"ff664db3-160a-4207-a543-9f43fff5e93a\") " pod="openshift-marketplace/community-operators-9zwp8" Dec 27 08:12:01 crc kubenswrapper[4934]: I1227 08:12:01.861495 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff664db3-160a-4207-a543-9f43fff5e93a-catalog-content\") pod \"community-operators-9zwp8\" (UID: \"ff664db3-160a-4207-a543-9f43fff5e93a\") " pod="openshift-marketplace/community-operators-9zwp8" Dec 27 08:12:01 crc kubenswrapper[4934]: I1227 08:12:01.861611 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff664db3-160a-4207-a543-9f43fff5e93a-utilities\") pod \"community-operators-9zwp8\" (UID: \"ff664db3-160a-4207-a543-9f43fff5e93a\") " pod="openshift-marketplace/community-operators-9zwp8" Dec 27 08:12:01 crc kubenswrapper[4934]: I1227 08:12:01.884982 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjzcb\" (UniqueName: \"kubernetes.io/projected/ff664db3-160a-4207-a543-9f43fff5e93a-kube-api-access-cjzcb\") pod \"community-operators-9zwp8\" (UID: \"ff664db3-160a-4207-a543-9f43fff5e93a\") " pod="openshift-marketplace/community-operators-9zwp8" Dec 27 08:12:01 crc kubenswrapper[4934]: I1227 08:12:01.922104 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9zwp8" Dec 27 08:12:02 crc kubenswrapper[4934]: I1227 08:12:02.464487 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9zwp8"] Dec 27 08:12:02 crc kubenswrapper[4934]: I1227 08:12:02.791646 4934 generic.go:334] "Generic (PLEG): container finished" podID="ff664db3-160a-4207-a543-9f43fff5e93a" containerID="d9171c9f5175ddc19f150eb49c8933788915bd0af81fdbfca2003f41b02fc40b" exitCode=0 Dec 27 08:12:02 crc kubenswrapper[4934]: I1227 08:12:02.791741 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zwp8" event={"ID":"ff664db3-160a-4207-a543-9f43fff5e93a","Type":"ContainerDied","Data":"d9171c9f5175ddc19f150eb49c8933788915bd0af81fdbfca2003f41b02fc40b"} Dec 27 08:12:02 crc kubenswrapper[4934]: I1227 08:12:02.791795 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zwp8" event={"ID":"ff664db3-160a-4207-a543-9f43fff5e93a","Type":"ContainerStarted","Data":"cda727f9ca93b50d21cebb002b839c15a1500cf3992703e682c07a85eb1a4755"} Dec 27 08:12:03 crc kubenswrapper[4934]: I1227 08:12:03.806907 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zwp8" event={"ID":"ff664db3-160a-4207-a543-9f43fff5e93a","Type":"ContainerStarted","Data":"27211ff9643517c7a7573b365584bbc343e785f9e75ea16f433c5bdb54ac1ce4"} Dec 27 08:12:06 crc kubenswrapper[4934]: I1227 08:12:06.847072 4934 generic.go:334] "Generic (PLEG): container finished" podID="ff664db3-160a-4207-a543-9f43fff5e93a" containerID="27211ff9643517c7a7573b365584bbc343e785f9e75ea16f433c5bdb54ac1ce4" exitCode=0 Dec 27 08:12:06 crc kubenswrapper[4934]: I1227 08:12:06.847268 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zwp8" event={"ID":"ff664db3-160a-4207-a543-9f43fff5e93a","Type":"ContainerDied","Data":"27211ff9643517c7a7573b365584bbc343e785f9e75ea16f433c5bdb54ac1ce4"} Dec 27 08:12:07 crc kubenswrapper[4934]: I1227 08:12:07.496296 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 27 08:12:07 crc kubenswrapper[4934]: I1227 08:12:07.861221 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zwp8" event={"ID":"ff664db3-160a-4207-a543-9f43fff5e93a","Type":"ContainerStarted","Data":"d35ede7144a5fcdd9f1708b1eba41fcf23c545afaf1ee78c510dc2d40384b947"} Dec 27 08:12:07 crc kubenswrapper[4934]: I1227 08:12:07.883036 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9zwp8" podStartSLOduration=2.333633497 podStartE2EDuration="6.883016681s" podCreationTimestamp="2025-12-27 08:12:01 +0000 UTC" firstStartedPulling="2025-12-27 08:12:02.79382197 +0000 UTC m=+1783.614262564" lastFinishedPulling="2025-12-27 08:12:07.343205124 +0000 UTC m=+1788.163645748" observedRunningTime="2025-12-27 08:12:07.881126114 +0000 UTC m=+1788.701566708" watchObservedRunningTime="2025-12-27 08:12:07.883016681 +0000 UTC m=+1788.703457275" Dec 27 08:12:11 crc kubenswrapper[4934]: I1227 08:12:11.922462 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9zwp8" Dec 27 08:12:11 crc kubenswrapper[4934]: I1227 08:12:11.922873 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9zwp8" Dec 27 08:12:12 crc kubenswrapper[4934]: I1227 08:12:12.016803 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9zwp8" Dec 27 08:12:12 crc kubenswrapper[4934]: I1227 08:12:12.468953 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:12:12 crc kubenswrapper[4934]: E1227 08:12:12.469659 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:12:13 crc kubenswrapper[4934]: I1227 08:12:13.010600 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9zwp8" Dec 27 08:12:13 crc kubenswrapper[4934]: I1227 08:12:13.073653 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9zwp8"] Dec 27 08:12:14 crc kubenswrapper[4934]: I1227 08:12:14.958998 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9zwp8" podUID="ff664db3-160a-4207-a543-9f43fff5e93a" containerName="registry-server" containerID="cri-o://d35ede7144a5fcdd9f1708b1eba41fcf23c545afaf1ee78c510dc2d40384b947" gracePeriod=2 Dec 27 08:12:15 crc kubenswrapper[4934]: I1227 08:12:15.578664 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9zwp8" Dec 27 08:12:15 crc kubenswrapper[4934]: I1227 08:12:15.697749 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjzcb\" (UniqueName: \"kubernetes.io/projected/ff664db3-160a-4207-a543-9f43fff5e93a-kube-api-access-cjzcb\") pod \"ff664db3-160a-4207-a543-9f43fff5e93a\" (UID: \"ff664db3-160a-4207-a543-9f43fff5e93a\") " Dec 27 08:12:15 crc kubenswrapper[4934]: I1227 08:12:15.697834 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff664db3-160a-4207-a543-9f43fff5e93a-utilities\") pod \"ff664db3-160a-4207-a543-9f43fff5e93a\" (UID: \"ff664db3-160a-4207-a543-9f43fff5e93a\") " Dec 27 08:12:15 crc kubenswrapper[4934]: I1227 08:12:15.698057 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff664db3-160a-4207-a543-9f43fff5e93a-catalog-content\") pod \"ff664db3-160a-4207-a543-9f43fff5e93a\" (UID: \"ff664db3-160a-4207-a543-9f43fff5e93a\") " Dec 27 08:12:15 crc kubenswrapper[4934]: I1227 08:12:15.701274 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff664db3-160a-4207-a543-9f43fff5e93a-utilities" (OuterVolumeSpecName: "utilities") pod "ff664db3-160a-4207-a543-9f43fff5e93a" (UID: "ff664db3-160a-4207-a543-9f43fff5e93a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:12:15 crc kubenswrapper[4934]: I1227 08:12:15.709350 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff664db3-160a-4207-a543-9f43fff5e93a-kube-api-access-cjzcb" (OuterVolumeSpecName: "kube-api-access-cjzcb") pod "ff664db3-160a-4207-a543-9f43fff5e93a" (UID: "ff664db3-160a-4207-a543-9f43fff5e93a"). InnerVolumeSpecName "kube-api-access-cjzcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:12:15 crc kubenswrapper[4934]: I1227 08:12:15.767301 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff664db3-160a-4207-a543-9f43fff5e93a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff664db3-160a-4207-a543-9f43fff5e93a" (UID: "ff664db3-160a-4207-a543-9f43fff5e93a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:12:15 crc kubenswrapper[4934]: I1227 08:12:15.802262 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff664db3-160a-4207-a543-9f43fff5e93a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 08:12:15 crc kubenswrapper[4934]: I1227 08:12:15.802309 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjzcb\" (UniqueName: \"kubernetes.io/projected/ff664db3-160a-4207-a543-9f43fff5e93a-kube-api-access-cjzcb\") on node \"crc\" DevicePath \"\"" Dec 27 08:12:15 crc kubenswrapper[4934]: I1227 08:12:15.802321 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff664db3-160a-4207-a543-9f43fff5e93a-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 08:12:15 crc kubenswrapper[4934]: I1227 08:12:15.976979 4934 generic.go:334] "Generic (PLEG): container finished" podID="ff664db3-160a-4207-a543-9f43fff5e93a" containerID="d35ede7144a5fcdd9f1708b1eba41fcf23c545afaf1ee78c510dc2d40384b947" exitCode=0 Dec 27 08:12:15 crc kubenswrapper[4934]: I1227 08:12:15.977058 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zwp8" event={"ID":"ff664db3-160a-4207-a543-9f43fff5e93a","Type":"ContainerDied","Data":"d35ede7144a5fcdd9f1708b1eba41fcf23c545afaf1ee78c510dc2d40384b947"} Dec 27 08:12:15 crc kubenswrapper[4934]: I1227 08:12:15.977143 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zwp8" event={"ID":"ff664db3-160a-4207-a543-9f43fff5e93a","Type":"ContainerDied","Data":"cda727f9ca93b50d21cebb002b839c15a1500cf3992703e682c07a85eb1a4755"} Dec 27 08:12:15 crc kubenswrapper[4934]: I1227 08:12:15.977185 4934 scope.go:117] "RemoveContainer" containerID="d35ede7144a5fcdd9f1708b1eba41fcf23c545afaf1ee78c510dc2d40384b947" Dec 27 08:12:15 crc kubenswrapper[4934]: I1227 08:12:15.978215 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9zwp8" Dec 27 08:12:16 crc kubenswrapper[4934]: I1227 08:12:16.019747 4934 scope.go:117] "RemoveContainer" containerID="27211ff9643517c7a7573b365584bbc343e785f9e75ea16f433c5bdb54ac1ce4" Dec 27 08:12:16 crc kubenswrapper[4934]: I1227 08:12:16.049464 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9zwp8"] Dec 27 08:12:16 crc kubenswrapper[4934]: I1227 08:12:16.061800 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9zwp8"] Dec 27 08:12:16 crc kubenswrapper[4934]: I1227 08:12:16.085140 4934 scope.go:117] "RemoveContainer" containerID="d9171c9f5175ddc19f150eb49c8933788915bd0af81fdbfca2003f41b02fc40b" Dec 27 08:12:16 crc kubenswrapper[4934]: I1227 08:12:16.142739 4934 scope.go:117] "RemoveContainer" containerID="d35ede7144a5fcdd9f1708b1eba41fcf23c545afaf1ee78c510dc2d40384b947" Dec 27 08:12:16 crc kubenswrapper[4934]: E1227 08:12:16.143242 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d35ede7144a5fcdd9f1708b1eba41fcf23c545afaf1ee78c510dc2d40384b947\": container with ID starting with d35ede7144a5fcdd9f1708b1eba41fcf23c545afaf1ee78c510dc2d40384b947 not found: ID does not exist" containerID="d35ede7144a5fcdd9f1708b1eba41fcf23c545afaf1ee78c510dc2d40384b947" Dec 27 08:12:16 crc kubenswrapper[4934]: I1227 08:12:16.143272 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d35ede7144a5fcdd9f1708b1eba41fcf23c545afaf1ee78c510dc2d40384b947"} err="failed to get container status \"d35ede7144a5fcdd9f1708b1eba41fcf23c545afaf1ee78c510dc2d40384b947\": rpc error: code = NotFound desc = could not find container \"d35ede7144a5fcdd9f1708b1eba41fcf23c545afaf1ee78c510dc2d40384b947\": container with ID starting with d35ede7144a5fcdd9f1708b1eba41fcf23c545afaf1ee78c510dc2d40384b947 not found: ID does not exist" Dec 27 08:12:16 crc kubenswrapper[4934]: I1227 08:12:16.143292 4934 scope.go:117] "RemoveContainer" containerID="27211ff9643517c7a7573b365584bbc343e785f9e75ea16f433c5bdb54ac1ce4" Dec 27 08:12:16 crc kubenswrapper[4934]: E1227 08:12:16.143608 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27211ff9643517c7a7573b365584bbc343e785f9e75ea16f433c5bdb54ac1ce4\": container with ID starting with 27211ff9643517c7a7573b365584bbc343e785f9e75ea16f433c5bdb54ac1ce4 not found: ID does not exist" containerID="27211ff9643517c7a7573b365584bbc343e785f9e75ea16f433c5bdb54ac1ce4" Dec 27 08:12:16 crc kubenswrapper[4934]: I1227 08:12:16.143631 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27211ff9643517c7a7573b365584bbc343e785f9e75ea16f433c5bdb54ac1ce4"} err="failed to get container status \"27211ff9643517c7a7573b365584bbc343e785f9e75ea16f433c5bdb54ac1ce4\": rpc error: code = NotFound desc = could not find container \"27211ff9643517c7a7573b365584bbc343e785f9e75ea16f433c5bdb54ac1ce4\": container with ID starting with 27211ff9643517c7a7573b365584bbc343e785f9e75ea16f433c5bdb54ac1ce4 not found: ID does not exist" Dec 27 08:12:16 crc kubenswrapper[4934]: I1227 08:12:16.143648 4934 scope.go:117] "RemoveContainer" containerID="d9171c9f5175ddc19f150eb49c8933788915bd0af81fdbfca2003f41b02fc40b" Dec 27 08:12:16 crc kubenswrapper[4934]: E1227 08:12:16.143855 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9171c9f5175ddc19f150eb49c8933788915bd0af81fdbfca2003f41b02fc40b\": container with ID starting with d9171c9f5175ddc19f150eb49c8933788915bd0af81fdbfca2003f41b02fc40b not found: ID does not exist" containerID="d9171c9f5175ddc19f150eb49c8933788915bd0af81fdbfca2003f41b02fc40b" Dec 27 08:12:16 crc kubenswrapper[4934]: I1227 08:12:16.143874 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9171c9f5175ddc19f150eb49c8933788915bd0af81fdbfca2003f41b02fc40b"} err="failed to get container status \"d9171c9f5175ddc19f150eb49c8933788915bd0af81fdbfca2003f41b02fc40b\": rpc error: code = NotFound desc = could not find container \"d9171c9f5175ddc19f150eb49c8933788915bd0af81fdbfca2003f41b02fc40b\": container with ID starting with d9171c9f5175ddc19f150eb49c8933788915bd0af81fdbfca2003f41b02fc40b not found: ID does not exist" Dec 27 08:12:17 crc kubenswrapper[4934]: I1227 08:12:17.493313 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff664db3-160a-4207-a543-9f43fff5e93a" path="/var/lib/kubelet/pods/ff664db3-160a-4207-a543-9f43fff5e93a/volumes" Dec 27 08:12:24 crc kubenswrapper[4934]: I1227 08:12:24.900273 4934 scope.go:117] "RemoveContainer" containerID="80405edd7bb455015027bd7125e6267a53d052b855fe164cffc53d0ef0acd2da" Dec 27 08:12:25 crc kubenswrapper[4934]: I1227 08:12:25.468872 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:12:25 crc kubenswrapper[4934]: E1227 08:12:25.470152 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:12:38 crc kubenswrapper[4934]: I1227 08:12:38.469249 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:12:38 crc kubenswrapper[4934]: E1227 08:12:38.469866 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:12:51 crc kubenswrapper[4934]: I1227 08:12:51.467826 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:12:51 crc kubenswrapper[4934]: E1227 08:12:51.469057 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:13:03 crc kubenswrapper[4934]: I1227 08:13:03.467991 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:13:03 crc kubenswrapper[4934]: E1227 08:13:03.470887 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:13:05 crc kubenswrapper[4934]: I1227 08:13:05.043713 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-29dwl"] Dec 27 08:13:05 crc kubenswrapper[4934]: I1227 08:13:05.056011 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-g2q82"] Dec 27 08:13:05 crc kubenswrapper[4934]: I1227 08:13:05.066302 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-29dwl"] Dec 27 08:13:05 crc kubenswrapper[4934]: I1227 08:13:05.075646 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-g2q82"] Dec 27 08:13:05 crc kubenswrapper[4934]: I1227 08:13:05.488380 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="956ddca3-cc80-4e88-8695-4d18f77c9292" path="/var/lib/kubelet/pods/956ddca3-cc80-4e88-8695-4d18f77c9292/volumes" Dec 27 08:13:05 crc kubenswrapper[4934]: I1227 08:13:05.490387 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0ade00b-01f5-493e-89ab-1925176e3bf0" path="/var/lib/kubelet/pods/c0ade00b-01f5-493e-89ab-1925176e3bf0/volumes" Dec 27 08:13:08 crc kubenswrapper[4934]: I1227 08:13:08.033995 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-wqrwd"] Dec 27 08:13:08 crc kubenswrapper[4934]: I1227 08:13:08.045841 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-wqrwd"] Dec 27 08:13:09 crc kubenswrapper[4934]: I1227 08:13:09.488684 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a71b8c67-af1b-4c61-9820-fad0c957733a" path="/var/lib/kubelet/pods/a71b8c67-af1b-4c61-9820-fad0c957733a/volumes" Dec 27 08:13:17 crc kubenswrapper[4934]: I1227 08:13:17.468143 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:13:17 crc kubenswrapper[4934]: I1227 08:13:17.812214 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"deb6fb2d2d95a934db9a65eebdf5e58dc64e13629a44e74765a4e64756d050bb"} Dec 27 08:13:20 crc kubenswrapper[4934]: I1227 08:13:20.040310 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-cgffw"] Dec 27 08:13:20 crc kubenswrapper[4934]: I1227 08:13:20.058306 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-cgffw"] Dec 27 08:13:21 crc kubenswrapper[4934]: I1227 08:13:21.044905 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-622b-account-create-rz2xx"] Dec 27 08:13:21 crc kubenswrapper[4934]: I1227 08:13:21.064031 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-622b-account-create-rz2xx"] Dec 27 08:13:21 crc kubenswrapper[4934]: I1227 08:13:21.094990 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-f752-account-create-mxw52"] Dec 27 08:13:21 crc kubenswrapper[4934]: I1227 08:13:21.112943 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-f752-account-create-mxw52"] Dec 27 08:13:21 crc kubenswrapper[4934]: I1227 08:13:21.127901 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-f9d8-account-create-dxnbx"] Dec 27 08:13:21 crc kubenswrapper[4934]: I1227 08:13:21.139587 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-f9d8-account-create-dxnbx"] Dec 27 08:13:21 crc kubenswrapper[4934]: I1227 08:13:21.488035 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e6e9101-e025-4566-833b-fd29fa46ab79" path="/var/lib/kubelet/pods/4e6e9101-e025-4566-833b-fd29fa46ab79/volumes" Dec 27 08:13:21 crc kubenswrapper[4934]: I1227 08:13:21.489764 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="766e151c-8962-4bd5-b174-4e381730ff6f" path="/var/lib/kubelet/pods/766e151c-8962-4bd5-b174-4e381730ff6f/volumes" Dec 27 08:13:21 crc kubenswrapper[4934]: I1227 08:13:21.491458 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ede7d3e-b268-4132-b5d3-70569e23f060" path="/var/lib/kubelet/pods/7ede7d3e-b268-4132-b5d3-70569e23f060/volumes" Dec 27 08:13:21 crc kubenswrapper[4934]: I1227 08:13:21.493235 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84b8ef67-1fd1-4d7c-82d3-cde56ca1468a" path="/var/lib/kubelet/pods/84b8ef67-1fd1-4d7c-82d3-cde56ca1468a/volumes" Dec 27 08:13:24 crc kubenswrapper[4934]: I1227 08:13:24.992059 4934 scope.go:117] "RemoveContainer" containerID="93f68cbe460b9e15249c70d01c06cbd2bdd7a3da61f34a0b42945ed88fab9f8f" Dec 27 08:13:25 crc kubenswrapper[4934]: I1227 08:13:25.042034 4934 scope.go:117] "RemoveContainer" containerID="c67d0c4914ff38adf00e96804612222913486319608a18a6c89859be0377e391" Dec 27 08:13:25 crc kubenswrapper[4934]: I1227 08:13:25.094176 4934 scope.go:117] "RemoveContainer" containerID="ddb9e9c22e3038ed6d0a692612f9934cb1e02c84c16773de0c65c293948fbc13" Dec 27 08:13:25 crc kubenswrapper[4934]: I1227 08:13:25.131350 4934 scope.go:117] "RemoveContainer" containerID="51323018708b2fb0bac023ea4fca5f9b434fc9be7f140175cd520dd278dc4336" Dec 27 08:13:25 crc kubenswrapper[4934]: I1227 08:13:25.227452 4934 scope.go:117] "RemoveContainer" containerID="236fa6ccd4c0ae3a62409338ff24ba79be6a75b377d52ba6570f2da9651436b3" Dec 27 08:13:25 crc kubenswrapper[4934]: I1227 08:13:25.276097 4934 scope.go:117] "RemoveContainer" containerID="0249d45e628cf667116fcfea30516924fefff2ec7c5ae9b97447a5e0d5cf5011" Dec 27 08:13:25 crc kubenswrapper[4934]: I1227 08:13:25.323454 4934 scope.go:117] "RemoveContainer" containerID="25069ee0e99208f73e7ad1aa0c872b35b7fddc1fe9590dda7a921889999be47d" Dec 27 08:13:25 crc kubenswrapper[4934]: I1227 08:13:25.368548 4934 scope.go:117] "RemoveContainer" containerID="d140ccfe6a91e2ba03a1ec5b8f9e5b1705101d5ce3274a3b70cf3771a3572cf2" Dec 27 08:13:25 crc kubenswrapper[4934]: I1227 08:13:25.401273 4934 scope.go:117] "RemoveContainer" containerID="9bdfc08d4542de65fa651c74626a5caf350ea01f0662434b7d89dc99e7088a05" Dec 27 08:13:26 crc kubenswrapper[4934]: I1227 08:13:26.977920 4934 generic.go:334] "Generic (PLEG): container finished" podID="68f2fdda-08a1-457f-b61e-c6d68ea69be3" containerID="14d7cfeda23d75c493a93c5ae393ee825df59024358ca93161c9539286658d13" exitCode=0 Dec 27 08:13:26 crc kubenswrapper[4934]: I1227 08:13:26.977992 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8" event={"ID":"68f2fdda-08a1-457f-b61e-c6d68ea69be3","Type":"ContainerDied","Data":"14d7cfeda23d75c493a93c5ae393ee825df59024358ca93161c9539286658d13"} Dec 27 08:13:28 crc kubenswrapper[4934]: I1227 08:13:28.071899 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-mbcsv"] Dec 27 08:13:28 crc kubenswrapper[4934]: I1227 08:13:28.085755 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-mbcsv"] Dec 27 08:13:28 crc kubenswrapper[4934]: I1227 08:13:28.584327 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8" Dec 27 08:13:28 crc kubenswrapper[4934]: I1227 08:13:28.661534 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68f2fdda-08a1-457f-b61e-c6d68ea69be3-ssh-key\") pod \"68f2fdda-08a1-457f-b61e-c6d68ea69be3\" (UID: \"68f2fdda-08a1-457f-b61e-c6d68ea69be3\") " Dec 27 08:13:28 crc kubenswrapper[4934]: I1227 08:13:28.661745 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8jf2\" (UniqueName: \"kubernetes.io/projected/68f2fdda-08a1-457f-b61e-c6d68ea69be3-kube-api-access-f8jf2\") pod \"68f2fdda-08a1-457f-b61e-c6d68ea69be3\" (UID: \"68f2fdda-08a1-457f-b61e-c6d68ea69be3\") " Dec 27 08:13:28 crc kubenswrapper[4934]: I1227 08:13:28.661919 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68f2fdda-08a1-457f-b61e-c6d68ea69be3-bootstrap-combined-ca-bundle\") pod \"68f2fdda-08a1-457f-b61e-c6d68ea69be3\" (UID: \"68f2fdda-08a1-457f-b61e-c6d68ea69be3\") " Dec 27 08:13:28 crc kubenswrapper[4934]: I1227 08:13:28.662204 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68f2fdda-08a1-457f-b61e-c6d68ea69be3-inventory\") pod \"68f2fdda-08a1-457f-b61e-c6d68ea69be3\" (UID: \"68f2fdda-08a1-457f-b61e-c6d68ea69be3\") " Dec 27 08:13:28 crc kubenswrapper[4934]: I1227 08:13:28.673310 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68f2fdda-08a1-457f-b61e-c6d68ea69be3-kube-api-access-f8jf2" (OuterVolumeSpecName: "kube-api-access-f8jf2") pod "68f2fdda-08a1-457f-b61e-c6d68ea69be3" (UID: "68f2fdda-08a1-457f-b61e-c6d68ea69be3"). InnerVolumeSpecName "kube-api-access-f8jf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:13:28 crc kubenswrapper[4934]: I1227 08:13:28.674537 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68f2fdda-08a1-457f-b61e-c6d68ea69be3-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "68f2fdda-08a1-457f-b61e-c6d68ea69be3" (UID: "68f2fdda-08a1-457f-b61e-c6d68ea69be3"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:13:28 crc kubenswrapper[4934]: I1227 08:13:28.714751 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68f2fdda-08a1-457f-b61e-c6d68ea69be3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "68f2fdda-08a1-457f-b61e-c6d68ea69be3" (UID: "68f2fdda-08a1-457f-b61e-c6d68ea69be3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:13:28 crc kubenswrapper[4934]: I1227 08:13:28.717194 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68f2fdda-08a1-457f-b61e-c6d68ea69be3-inventory" (OuterVolumeSpecName: "inventory") pod "68f2fdda-08a1-457f-b61e-c6d68ea69be3" (UID: "68f2fdda-08a1-457f-b61e-c6d68ea69be3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:13:28 crc kubenswrapper[4934]: I1227 08:13:28.764762 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8jf2\" (UniqueName: \"kubernetes.io/projected/68f2fdda-08a1-457f-b61e-c6d68ea69be3-kube-api-access-f8jf2\") on node \"crc\" DevicePath \"\"" Dec 27 08:13:28 crc kubenswrapper[4934]: I1227 08:13:28.764818 4934 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68f2fdda-08a1-457f-b61e-c6d68ea69be3-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:13:28 crc kubenswrapper[4934]: I1227 08:13:28.764831 4934 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68f2fdda-08a1-457f-b61e-c6d68ea69be3-inventory\") on node \"crc\" DevicePath \"\"" Dec 27 08:13:28 crc kubenswrapper[4934]: I1227 08:13:28.764839 4934 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/68f2fdda-08a1-457f-b61e-c6d68ea69be3-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.023650 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8" event={"ID":"68f2fdda-08a1-457f-b61e-c6d68ea69be3","Type":"ContainerDied","Data":"51d7c6c5e1379aa66eb84e9cf98f47745c29e81607b2d537cede28de3c4b0c35"} Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.023716 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51d7c6c5e1379aa66eb84e9cf98f47745c29e81607b2d537cede28de3c4b0c35" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.023720 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.200771 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5"] Dec 27 08:13:29 crc kubenswrapper[4934]: E1227 08:13:29.201314 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68f2fdda-08a1-457f-b61e-c6d68ea69be3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.201330 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="68f2fdda-08a1-457f-b61e-c6d68ea69be3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 27 08:13:29 crc kubenswrapper[4934]: E1227 08:13:29.201350 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff664db3-160a-4207-a543-9f43fff5e93a" containerName="extract-content" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.201358 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff664db3-160a-4207-a543-9f43fff5e93a" containerName="extract-content" Dec 27 08:13:29 crc kubenswrapper[4934]: E1227 08:13:29.201384 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff664db3-160a-4207-a543-9f43fff5e93a" containerName="extract-utilities" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.201392 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff664db3-160a-4207-a543-9f43fff5e93a" containerName="extract-utilities" Dec 27 08:13:29 crc kubenswrapper[4934]: E1227 08:13:29.201440 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff664db3-160a-4207-a543-9f43fff5e93a" containerName="registry-server" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.201448 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff664db3-160a-4207-a543-9f43fff5e93a" containerName="registry-server" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.201753 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="68f2fdda-08a1-457f-b61e-c6d68ea69be3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.201772 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff664db3-160a-4207-a543-9f43fff5e93a" containerName="registry-server" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.202746 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.205786 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.206011 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.206208 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.206432 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-98ddn" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.239631 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5"] Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.278165 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6a50c262-1a91-4dad-91ec-6530d981415a-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5\" (UID: \"6a50c262-1a91-4dad-91ec-6530d981415a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.278278 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6a50c262-1a91-4dad-91ec-6530d981415a-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5\" (UID: \"6a50c262-1a91-4dad-91ec-6530d981415a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.278371 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjz8t\" (UniqueName: \"kubernetes.io/projected/6a50c262-1a91-4dad-91ec-6530d981415a-kube-api-access-jjz8t\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5\" (UID: \"6a50c262-1a91-4dad-91ec-6530d981415a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.380566 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6a50c262-1a91-4dad-91ec-6530d981415a-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5\" (UID: \"6a50c262-1a91-4dad-91ec-6530d981415a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.380703 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6a50c262-1a91-4dad-91ec-6530d981415a-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5\" (UID: \"6a50c262-1a91-4dad-91ec-6530d981415a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.380770 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjz8t\" (UniqueName: \"kubernetes.io/projected/6a50c262-1a91-4dad-91ec-6530d981415a-kube-api-access-jjz8t\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5\" (UID: \"6a50c262-1a91-4dad-91ec-6530d981415a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.386759 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6a50c262-1a91-4dad-91ec-6530d981415a-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5\" (UID: \"6a50c262-1a91-4dad-91ec-6530d981415a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.387601 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6a50c262-1a91-4dad-91ec-6530d981415a-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5\" (UID: \"6a50c262-1a91-4dad-91ec-6530d981415a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.413811 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjz8t\" (UniqueName: \"kubernetes.io/projected/6a50c262-1a91-4dad-91ec-6530d981415a-kube-api-access-jjz8t\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5\" (UID: \"6a50c262-1a91-4dad-91ec-6530d981415a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.484581 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59b9ab13-2205-48a4-8a33-9d85f7de7d04" path="/var/lib/kubelet/pods/59b9ab13-2205-48a4-8a33-9d85f7de7d04/volumes" Dec 27 08:13:29 crc kubenswrapper[4934]: I1227 08:13:29.524457 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5" Dec 27 08:13:30 crc kubenswrapper[4934]: I1227 08:13:30.097750 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5"] Dec 27 08:13:31 crc kubenswrapper[4934]: I1227 08:13:31.047269 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-090b-account-create-89jvs"] Dec 27 08:13:31 crc kubenswrapper[4934]: I1227 08:13:31.053737 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5" event={"ID":"6a50c262-1a91-4dad-91ec-6530d981415a","Type":"ContainerStarted","Data":"fb5f99f49dc96c4b749db670a3507b9ae85ae0f930f4dc9df65bfad0fbb6070d"} Dec 27 08:13:31 crc kubenswrapper[4934]: I1227 08:13:31.053791 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5" event={"ID":"6a50c262-1a91-4dad-91ec-6530d981415a","Type":"ContainerStarted","Data":"259a6976c0342102cdd656ddce3e8b7ae4bcd91eef21326c8f9dac4a2bb40195"} Dec 27 08:13:31 crc kubenswrapper[4934]: I1227 08:13:31.065711 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-090b-account-create-89jvs"] Dec 27 08:13:31 crc kubenswrapper[4934]: I1227 08:13:31.077787 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5" podStartSLOduration=1.629741739 podStartE2EDuration="2.077770527s" podCreationTimestamp="2025-12-27 08:13:29 +0000 UTC" firstStartedPulling="2025-12-27 08:13:30.107139388 +0000 UTC m=+1870.927580002" lastFinishedPulling="2025-12-27 08:13:30.555168186 +0000 UTC m=+1871.375608790" observedRunningTime="2025-12-27 08:13:31.068009646 +0000 UTC m=+1871.888450290" watchObservedRunningTime="2025-12-27 08:13:31.077770527 +0000 UTC m=+1871.898211121" Dec 27 08:13:31 crc kubenswrapper[4934]: I1227 08:13:31.484534 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bb54f3d-f01e-445f-beee-2a1f33537008" path="/var/lib/kubelet/pods/9bb54f3d-f01e-445f-beee-2a1f33537008/volumes" Dec 27 08:13:38 crc kubenswrapper[4934]: I1227 08:13:38.042288 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-bef3-account-create-qm6cg"] Dec 27 08:13:38 crc kubenswrapper[4934]: I1227 08:13:38.052833 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-bef3-account-create-qm6cg"] Dec 27 08:13:39 crc kubenswrapper[4934]: I1227 08:13:39.491310 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9f93658-b339-4655-841b-bebfba9ad231" path="/var/lib/kubelet/pods/c9f93658-b339-4655-841b-bebfba9ad231/volumes" Dec 27 08:13:58 crc kubenswrapper[4934]: I1227 08:13:58.043072 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-dhmx4"] Dec 27 08:13:58 crc kubenswrapper[4934]: I1227 08:13:58.063388 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-vcjnz"] Dec 27 08:13:58 crc kubenswrapper[4934]: I1227 08:13:58.078652 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-dhmx4"] Dec 27 08:13:58 crc kubenswrapper[4934]: I1227 08:13:58.089168 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-vcjnz"] Dec 27 08:13:58 crc kubenswrapper[4934]: I1227 08:13:58.099678 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-858rb"] Dec 27 08:13:58 crc kubenswrapper[4934]: I1227 08:13:58.109629 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-92xmf"] Dec 27 08:13:58 crc kubenswrapper[4934]: I1227 08:13:58.120689 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-858rb"] Dec 27 08:13:58 crc kubenswrapper[4934]: I1227 08:13:58.131345 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-92xmf"] Dec 27 08:13:59 crc kubenswrapper[4934]: I1227 08:13:59.485255 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06fd3b33-8755-4e16-93df-6ca70836a8f7" path="/var/lib/kubelet/pods/06fd3b33-8755-4e16-93df-6ca70836a8f7/volumes" Dec 27 08:13:59 crc kubenswrapper[4934]: I1227 08:13:59.486516 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47797090-7e7d-40ba-bf3f-a43d18af9283" path="/var/lib/kubelet/pods/47797090-7e7d-40ba-bf3f-a43d18af9283/volumes" Dec 27 08:13:59 crc kubenswrapper[4934]: I1227 08:13:59.489117 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4aa74e5-8fd8-45e5-907d-e95f8bf11d37" path="/var/lib/kubelet/pods/a4aa74e5-8fd8-45e5-907d-e95f8bf11d37/volumes" Dec 27 08:13:59 crc kubenswrapper[4934]: I1227 08:13:59.489848 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d745a073-8cd4-463c-ae78-53185e910777" path="/var/lib/kubelet/pods/d745a073-8cd4-463c-ae78-53185e910777/volumes" Dec 27 08:14:06 crc kubenswrapper[4934]: I1227 08:14:06.062387 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-cgptv"] Dec 27 08:14:06 crc kubenswrapper[4934]: I1227 08:14:06.078282 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-cgptv"] Dec 27 08:14:07 crc kubenswrapper[4934]: I1227 08:14:07.481201 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c811bf9-1bb4-4f9b-8967-9a67bb94c520" path="/var/lib/kubelet/pods/5c811bf9-1bb4-4f9b-8967-9a67bb94c520/volumes" Dec 27 08:14:13 crc kubenswrapper[4934]: I1227 08:14:13.045829 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-e80b-account-create-fvzrf"] Dec 27 08:14:13 crc kubenswrapper[4934]: I1227 08:14:13.061654 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-e80b-account-create-fvzrf"] Dec 27 08:14:13 crc kubenswrapper[4934]: I1227 08:14:13.491251 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3489623-db32-44b7-9201-83a6f7603021" path="/var/lib/kubelet/pods/e3489623-db32-44b7-9201-83a6f7603021/volumes" Dec 27 08:14:15 crc kubenswrapper[4934]: I1227 08:14:15.031546 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-e74b-account-create-dwdlx"] Dec 27 08:14:15 crc kubenswrapper[4934]: I1227 08:14:15.041475 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-e74b-account-create-dwdlx"] Dec 27 08:14:15 crc kubenswrapper[4934]: I1227 08:14:15.481008 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c23056c-4932-4373-9e14-c3f93989eb7f" path="/var/lib/kubelet/pods/2c23056c-4932-4373-9e14-c3f93989eb7f/volumes" Dec 27 08:14:16 crc kubenswrapper[4934]: I1227 08:14:16.051428 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-a9fc-account-create-px94p"] Dec 27 08:14:16 crc kubenswrapper[4934]: I1227 08:14:16.065267 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-e4e3-account-create-ssb4t"] Dec 27 08:14:16 crc kubenswrapper[4934]: I1227 08:14:16.077847 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-a9fc-account-create-px94p"] Dec 27 08:14:16 crc kubenswrapper[4934]: I1227 08:14:16.088643 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-e4e3-account-create-ssb4t"] Dec 27 08:14:17 crc kubenswrapper[4934]: I1227 08:14:17.494057 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0328afdb-dc17-4fa6-9d22-44dca770b879" path="/var/lib/kubelet/pods/0328afdb-dc17-4fa6-9d22-44dca770b879/volumes" Dec 27 08:14:17 crc kubenswrapper[4934]: I1227 08:14:17.496641 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1443d960-8241-4008-9a7a-7d32a158dd32" path="/var/lib/kubelet/pods/1443d960-8241-4008-9a7a-7d32a158dd32/volumes" Dec 27 08:14:18 crc kubenswrapper[4934]: I1227 08:14:18.052609 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-6vlxg"] Dec 27 08:14:18 crc kubenswrapper[4934]: I1227 08:14:18.078735 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-6vlxg"] Dec 27 08:14:19 crc kubenswrapper[4934]: I1227 08:14:19.488045 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94526b19-793f-4ec9-9a80-8a7d4f8cab87" path="/var/lib/kubelet/pods/94526b19-793f-4ec9-9a80-8a7d4f8cab87/volumes" Dec 27 08:14:23 crc kubenswrapper[4934]: I1227 08:14:23.037238 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-vkcbn"] Dec 27 08:14:23 crc kubenswrapper[4934]: I1227 08:14:23.055037 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-vkcbn"] Dec 27 08:14:23 crc kubenswrapper[4934]: I1227 08:14:23.492372 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7e0de58-4866-4ba7-9a80-7bce4eb52b97" path="/var/lib/kubelet/pods/f7e0de58-4866-4ba7-9a80-7bce4eb52b97/volumes" Dec 27 08:14:25 crc kubenswrapper[4934]: I1227 08:14:25.656079 4934 scope.go:117] "RemoveContainer" containerID="39b700197bd5740bb3755b1a69f4fa0e9e21e689c54f2288a3f7d636931af969" Dec 27 08:14:25 crc kubenswrapper[4934]: I1227 08:14:25.701691 4934 scope.go:117] "RemoveContainer" containerID="f82dec17909b52ec75d630a98ad35e72ac5164071010cf7cd387118c867dc70d" Dec 27 08:14:25 crc kubenswrapper[4934]: I1227 08:14:25.767851 4934 scope.go:117] "RemoveContainer" containerID="afb07e08e8bedb3662ba63846f90cd9ae80ee3a32b160bb54e44d523509b36b8" Dec 27 08:14:25 crc kubenswrapper[4934]: I1227 08:14:25.856566 4934 scope.go:117] "RemoveContainer" containerID="6edd9147c9e2c957f7c485e8994dc93a28028e64b952d232a8c9a535692b0117" Dec 27 08:14:25 crc kubenswrapper[4934]: I1227 08:14:25.913511 4934 scope.go:117] "RemoveContainer" containerID="9c7a962c5fbfa08d7139f70735d236be10782026de00fe694b1949d1d850814b" Dec 27 08:14:25 crc kubenswrapper[4934]: I1227 08:14:25.977885 4934 scope.go:117] "RemoveContainer" containerID="3f946b91baf02cf6d91daa86547ddb8ec839e63d7906f28412e858239ccaec2d" Dec 27 08:14:26 crc kubenswrapper[4934]: I1227 08:14:26.039853 4934 scope.go:117] "RemoveContainer" containerID="96d7f9fd2a893d377ce92367a2ddadd2206dbd691c9c4f16919c7d3017c33831" Dec 27 08:14:26 crc kubenswrapper[4934]: I1227 08:14:26.063331 4934 scope.go:117] "RemoveContainer" containerID="ec3f9d839dd4caf8943fdebc447b00e343dfdebc72e74bb103a9ad1cbd5d5875" Dec 27 08:14:26 crc kubenswrapper[4934]: I1227 08:14:26.098555 4934 scope.go:117] "RemoveContainer" containerID="cf486a099c077f6365252affac7ca7a4613758c7406362dc41872cda5fa53012" Dec 27 08:14:26 crc kubenswrapper[4934]: I1227 08:14:26.127540 4934 scope.go:117] "RemoveContainer" containerID="c3f86223e2285fd2790d23cc1157f81d5862840766502576d5b20ca6aac68bd2" Dec 27 08:14:26 crc kubenswrapper[4934]: I1227 08:14:26.155172 4934 scope.go:117] "RemoveContainer" containerID="afb8cdf9e621a319d8617a7c959615983ed7dacecc069a07889905c69202ed0d" Dec 27 08:14:26 crc kubenswrapper[4934]: I1227 08:14:26.189341 4934 scope.go:117] "RemoveContainer" containerID="6f9daf1e58f583c4f515329a771747e6c093a7c929ef3cdd821143d93d088151" Dec 27 08:14:26 crc kubenswrapper[4934]: I1227 08:14:26.215412 4934 scope.go:117] "RemoveContainer" containerID="58eec049f82987e1a6707f16ad161c530040c78d71a3c2b424895f1b1549ea3c" Dec 27 08:14:26 crc kubenswrapper[4934]: I1227 08:14:26.235964 4934 scope.go:117] "RemoveContainer" containerID="f33426463ede301b4ff1ea7a6f26b88450efde0a625ec9eda1972bc388900f56" Dec 27 08:14:35 crc kubenswrapper[4934]: I1227 08:14:35.056659 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-v7n52"] Dec 27 08:14:35 crc kubenswrapper[4934]: I1227 08:14:35.067456 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-v7n52"] Dec 27 08:14:35 crc kubenswrapper[4934]: I1227 08:14:35.485073 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77f95c7a-ce32-4353-8d7f-c5710d0b0057" path="/var/lib/kubelet/pods/77f95c7a-ce32-4353-8d7f-c5710d0b0057/volumes" Dec 27 08:14:55 crc kubenswrapper[4934]: I1227 08:14:55.055557 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-zpfs9"] Dec 27 08:14:55 crc kubenswrapper[4934]: I1227 08:14:55.067125 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-zpfs9"] Dec 27 08:14:55 crc kubenswrapper[4934]: I1227 08:14:55.480974 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38b7d77d-aea4-43e4-b5a9-4021d2562e4c" path="/var/lib/kubelet/pods/38b7d77d-aea4-43e4-b5a9-4021d2562e4c/volumes" Dec 27 08:15:00 crc kubenswrapper[4934]: I1227 08:15:00.170454 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c"] Dec 27 08:15:00 crc kubenswrapper[4934]: I1227 08:15:00.173851 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c" Dec 27 08:15:00 crc kubenswrapper[4934]: I1227 08:15:00.177334 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 27 08:15:00 crc kubenswrapper[4934]: I1227 08:15:00.178128 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 27 08:15:00 crc kubenswrapper[4934]: I1227 08:15:00.184517 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c"] Dec 27 08:15:00 crc kubenswrapper[4934]: I1227 08:15:00.315363 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1457a36-6d7f-4770-bca1-d7bc5919bb4c-config-volume\") pod \"collect-profiles-29447055-vg29c\" (UID: \"d1457a36-6d7f-4770-bca1-d7bc5919bb4c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c" Dec 27 08:15:00 crc kubenswrapper[4934]: I1227 08:15:00.315474 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d1457a36-6d7f-4770-bca1-d7bc5919bb4c-secret-volume\") pod \"collect-profiles-29447055-vg29c\" (UID: \"d1457a36-6d7f-4770-bca1-d7bc5919bb4c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c" Dec 27 08:15:00 crc kubenswrapper[4934]: I1227 08:15:00.315638 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm9p6\" (UniqueName: \"kubernetes.io/projected/d1457a36-6d7f-4770-bca1-d7bc5919bb4c-kube-api-access-wm9p6\") pod \"collect-profiles-29447055-vg29c\" (UID: \"d1457a36-6d7f-4770-bca1-d7bc5919bb4c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c" Dec 27 08:15:00 crc kubenswrapper[4934]: I1227 08:15:00.417808 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1457a36-6d7f-4770-bca1-d7bc5919bb4c-config-volume\") pod \"collect-profiles-29447055-vg29c\" (UID: \"d1457a36-6d7f-4770-bca1-d7bc5919bb4c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c" Dec 27 08:15:00 crc kubenswrapper[4934]: I1227 08:15:00.417912 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d1457a36-6d7f-4770-bca1-d7bc5919bb4c-secret-volume\") pod \"collect-profiles-29447055-vg29c\" (UID: \"d1457a36-6d7f-4770-bca1-d7bc5919bb4c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c" Dec 27 08:15:00 crc kubenswrapper[4934]: I1227 08:15:00.418011 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm9p6\" (UniqueName: \"kubernetes.io/projected/d1457a36-6d7f-4770-bca1-d7bc5919bb4c-kube-api-access-wm9p6\") pod \"collect-profiles-29447055-vg29c\" (UID: \"d1457a36-6d7f-4770-bca1-d7bc5919bb4c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c" Dec 27 08:15:00 crc kubenswrapper[4934]: I1227 08:15:00.420147 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1457a36-6d7f-4770-bca1-d7bc5919bb4c-config-volume\") pod \"collect-profiles-29447055-vg29c\" (UID: \"d1457a36-6d7f-4770-bca1-d7bc5919bb4c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c" Dec 27 08:15:00 crc kubenswrapper[4934]: I1227 08:15:00.427771 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d1457a36-6d7f-4770-bca1-d7bc5919bb4c-secret-volume\") pod \"collect-profiles-29447055-vg29c\" (UID: \"d1457a36-6d7f-4770-bca1-d7bc5919bb4c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c" Dec 27 08:15:00 crc kubenswrapper[4934]: I1227 08:15:00.439373 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm9p6\" (UniqueName: \"kubernetes.io/projected/d1457a36-6d7f-4770-bca1-d7bc5919bb4c-kube-api-access-wm9p6\") pod \"collect-profiles-29447055-vg29c\" (UID: \"d1457a36-6d7f-4770-bca1-d7bc5919bb4c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c" Dec 27 08:15:00 crc kubenswrapper[4934]: I1227 08:15:00.561362 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c" Dec 27 08:15:01 crc kubenswrapper[4934]: I1227 08:15:01.018964 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c"] Dec 27 08:15:01 crc kubenswrapper[4934]: I1227 08:15:01.345731 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c" event={"ID":"d1457a36-6d7f-4770-bca1-d7bc5919bb4c","Type":"ContainerStarted","Data":"d6f67af5a2fc359f79dffaecc06c0e92449e263cd574162ecfcec72d08c18736"} Dec 27 08:15:01 crc kubenswrapper[4934]: I1227 08:15:01.345787 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c" event={"ID":"d1457a36-6d7f-4770-bca1-d7bc5919bb4c","Type":"ContainerStarted","Data":"0e413995078fba4f4f2981a334b8afb27470ae39e782cdcec3dbbecd237becfa"} Dec 27 08:15:01 crc kubenswrapper[4934]: I1227 08:15:01.369374 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c" podStartSLOduration=1.36935495 podStartE2EDuration="1.36935495s" podCreationTimestamp="2025-12-27 08:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:15:01.367058093 +0000 UTC m=+1962.187498707" watchObservedRunningTime="2025-12-27 08:15:01.36935495 +0000 UTC m=+1962.189795544" Dec 27 08:15:02 crc kubenswrapper[4934]: I1227 08:15:02.364251 4934 generic.go:334] "Generic (PLEG): container finished" podID="d1457a36-6d7f-4770-bca1-d7bc5919bb4c" containerID="d6f67af5a2fc359f79dffaecc06c0e92449e263cd574162ecfcec72d08c18736" exitCode=0 Dec 27 08:15:02 crc kubenswrapper[4934]: I1227 08:15:02.364381 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c" event={"ID":"d1457a36-6d7f-4770-bca1-d7bc5919bb4c","Type":"ContainerDied","Data":"d6f67af5a2fc359f79dffaecc06c0e92449e263cd574162ecfcec72d08c18736"} Dec 27 08:15:03 crc kubenswrapper[4934]: I1227 08:15:03.828688 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c" Dec 27 08:15:03 crc kubenswrapper[4934]: I1227 08:15:03.946246 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wm9p6\" (UniqueName: \"kubernetes.io/projected/d1457a36-6d7f-4770-bca1-d7bc5919bb4c-kube-api-access-wm9p6\") pod \"d1457a36-6d7f-4770-bca1-d7bc5919bb4c\" (UID: \"d1457a36-6d7f-4770-bca1-d7bc5919bb4c\") " Dec 27 08:15:03 crc kubenswrapper[4934]: I1227 08:15:03.946518 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d1457a36-6d7f-4770-bca1-d7bc5919bb4c-secret-volume\") pod \"d1457a36-6d7f-4770-bca1-d7bc5919bb4c\" (UID: \"d1457a36-6d7f-4770-bca1-d7bc5919bb4c\") " Dec 27 08:15:03 crc kubenswrapper[4934]: I1227 08:15:03.946616 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1457a36-6d7f-4770-bca1-d7bc5919bb4c-config-volume\") pod \"d1457a36-6d7f-4770-bca1-d7bc5919bb4c\" (UID: \"d1457a36-6d7f-4770-bca1-d7bc5919bb4c\") " Dec 27 08:15:03 crc kubenswrapper[4934]: I1227 08:15:03.947272 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1457a36-6d7f-4770-bca1-d7bc5919bb4c-config-volume" (OuterVolumeSpecName: "config-volume") pod "d1457a36-6d7f-4770-bca1-d7bc5919bb4c" (UID: "d1457a36-6d7f-4770-bca1-d7bc5919bb4c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:15:03 crc kubenswrapper[4934]: I1227 08:15:03.952560 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1457a36-6d7f-4770-bca1-d7bc5919bb4c-kube-api-access-wm9p6" (OuterVolumeSpecName: "kube-api-access-wm9p6") pod "d1457a36-6d7f-4770-bca1-d7bc5919bb4c" (UID: "d1457a36-6d7f-4770-bca1-d7bc5919bb4c"). InnerVolumeSpecName "kube-api-access-wm9p6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:15:03 crc kubenswrapper[4934]: I1227 08:15:03.952969 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1457a36-6d7f-4770-bca1-d7bc5919bb4c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d1457a36-6d7f-4770-bca1-d7bc5919bb4c" (UID: "d1457a36-6d7f-4770-bca1-d7bc5919bb4c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:15:04 crc kubenswrapper[4934]: I1227 08:15:04.049472 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wm9p6\" (UniqueName: \"kubernetes.io/projected/d1457a36-6d7f-4770-bca1-d7bc5919bb4c-kube-api-access-wm9p6\") on node \"crc\" DevicePath \"\"" Dec 27 08:15:04 crc kubenswrapper[4934]: I1227 08:15:04.049509 4934 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d1457a36-6d7f-4770-bca1-d7bc5919bb4c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 27 08:15:04 crc kubenswrapper[4934]: I1227 08:15:04.049519 4934 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1457a36-6d7f-4770-bca1-d7bc5919bb4c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 27 08:15:04 crc kubenswrapper[4934]: I1227 08:15:04.398015 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c" event={"ID":"d1457a36-6d7f-4770-bca1-d7bc5919bb4c","Type":"ContainerDied","Data":"0e413995078fba4f4f2981a334b8afb27470ae39e782cdcec3dbbecd237becfa"} Dec 27 08:15:04 crc kubenswrapper[4934]: I1227 08:15:04.398339 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e413995078fba4f4f2981a334b8afb27470ae39e782cdcec3dbbecd237becfa" Dec 27 08:15:04 crc kubenswrapper[4934]: I1227 08:15:04.398148 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c" Dec 27 08:15:04 crc kubenswrapper[4934]: I1227 08:15:04.509914 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj"] Dec 27 08:15:04 crc kubenswrapper[4934]: I1227 08:15:04.531483 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447010-fl8rj"] Dec 27 08:15:05 crc kubenswrapper[4934]: I1227 08:15:05.484730 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce5a8998-bb68-4764-aa7d-a47ec23c4514" path="/var/lib/kubelet/pods/ce5a8998-bb68-4764-aa7d-a47ec23c4514/volumes" Dec 27 08:15:16 crc kubenswrapper[4934]: I1227 08:15:16.043596 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-f745s"] Dec 27 08:15:16 crc kubenswrapper[4934]: I1227 08:15:16.058262 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-f745s"] Dec 27 08:15:17 crc kubenswrapper[4934]: I1227 08:15:17.495280 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ef9a635-4d25-46d5-aa28-6c75f3cf8af4" path="/var/lib/kubelet/pods/5ef9a635-4d25-46d5-aa28-6c75f3cf8af4/volumes" Dec 27 08:15:20 crc kubenswrapper[4934]: I1227 08:15:20.050350 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-rrj89"] Dec 27 08:15:20 crc kubenswrapper[4934]: I1227 08:15:20.069394 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-rrj89"] Dec 27 08:15:21 crc kubenswrapper[4934]: I1227 08:15:21.483718 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72d0ed22-31be-4048-b031-139856c584c5" path="/var/lib/kubelet/pods/72d0ed22-31be-4048-b031-139856c584c5/volumes" Dec 27 08:15:26 crc kubenswrapper[4934]: I1227 08:15:26.625903 4934 scope.go:117] "RemoveContainer" containerID="f3f6afb85dd38d0d774bd8590ac099a7416e9ef26fa17b44420a18193ffcdcae" Dec 27 08:15:26 crc kubenswrapper[4934]: I1227 08:15:26.672640 4934 scope.go:117] "RemoveContainer" containerID="dd3aa6e013a5c434c3c2c70132d259a4f0966aee7c14cdfa3dc77dd5572fc961" Dec 27 08:15:26 crc kubenswrapper[4934]: I1227 08:15:26.754152 4934 scope.go:117] "RemoveContainer" containerID="7ee08fd9f1536658fe694fdf3478f7814115d9d18792f1ac758353bb2affd864" Dec 27 08:15:26 crc kubenswrapper[4934]: I1227 08:15:26.818687 4934 scope.go:117] "RemoveContainer" containerID="e4ca73aad4d3c626492002698fb4f4abfe8d862fe00362c088cb7f52b523eac4" Dec 27 08:15:26 crc kubenswrapper[4934]: I1227 08:15:26.881461 4934 scope.go:117] "RemoveContainer" containerID="cce8332487d739cc4965e08f8bea71277301b3c77c45c57b3a9fb5cbd3d65e86" Dec 27 08:15:40 crc kubenswrapper[4934]: I1227 08:15:40.887306 4934 generic.go:334] "Generic (PLEG): container finished" podID="6a50c262-1a91-4dad-91ec-6530d981415a" containerID="fb5f99f49dc96c4b749db670a3507b9ae85ae0f930f4dc9df65bfad0fbb6070d" exitCode=0 Dec 27 08:15:40 crc kubenswrapper[4934]: I1227 08:15:40.887396 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5" event={"ID":"6a50c262-1a91-4dad-91ec-6530d981415a","Type":"ContainerDied","Data":"fb5f99f49dc96c4b749db670a3507b9ae85ae0f930f4dc9df65bfad0fbb6070d"} Dec 27 08:15:42 crc kubenswrapper[4934]: I1227 08:15:42.484112 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5" Dec 27 08:15:42 crc kubenswrapper[4934]: I1227 08:15:42.574862 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjz8t\" (UniqueName: \"kubernetes.io/projected/6a50c262-1a91-4dad-91ec-6530d981415a-kube-api-access-jjz8t\") pod \"6a50c262-1a91-4dad-91ec-6530d981415a\" (UID: \"6a50c262-1a91-4dad-91ec-6530d981415a\") " Dec 27 08:15:42 crc kubenswrapper[4934]: I1227 08:15:42.574947 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6a50c262-1a91-4dad-91ec-6530d981415a-inventory\") pod \"6a50c262-1a91-4dad-91ec-6530d981415a\" (UID: \"6a50c262-1a91-4dad-91ec-6530d981415a\") " Dec 27 08:15:42 crc kubenswrapper[4934]: I1227 08:15:42.574997 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6a50c262-1a91-4dad-91ec-6530d981415a-ssh-key\") pod \"6a50c262-1a91-4dad-91ec-6530d981415a\" (UID: \"6a50c262-1a91-4dad-91ec-6530d981415a\") " Dec 27 08:15:42 crc kubenswrapper[4934]: I1227 08:15:42.582137 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a50c262-1a91-4dad-91ec-6530d981415a-kube-api-access-jjz8t" (OuterVolumeSpecName: "kube-api-access-jjz8t") pod "6a50c262-1a91-4dad-91ec-6530d981415a" (UID: "6a50c262-1a91-4dad-91ec-6530d981415a"). InnerVolumeSpecName "kube-api-access-jjz8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:15:42 crc kubenswrapper[4934]: I1227 08:15:42.639436 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a50c262-1a91-4dad-91ec-6530d981415a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6a50c262-1a91-4dad-91ec-6530d981415a" (UID: "6a50c262-1a91-4dad-91ec-6530d981415a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:15:42 crc kubenswrapper[4934]: I1227 08:15:42.639472 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a50c262-1a91-4dad-91ec-6530d981415a-inventory" (OuterVolumeSpecName: "inventory") pod "6a50c262-1a91-4dad-91ec-6530d981415a" (UID: "6a50c262-1a91-4dad-91ec-6530d981415a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:15:42 crc kubenswrapper[4934]: I1227 08:15:42.678829 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjz8t\" (UniqueName: \"kubernetes.io/projected/6a50c262-1a91-4dad-91ec-6530d981415a-kube-api-access-jjz8t\") on node \"crc\" DevicePath \"\"" Dec 27 08:15:42 crc kubenswrapper[4934]: I1227 08:15:42.678880 4934 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6a50c262-1a91-4dad-91ec-6530d981415a-inventory\") on node \"crc\" DevicePath \"\"" Dec 27 08:15:42 crc kubenswrapper[4934]: I1227 08:15:42.678899 4934 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6a50c262-1a91-4dad-91ec-6530d981415a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 27 08:15:42 crc kubenswrapper[4934]: I1227 08:15:42.913614 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5" event={"ID":"6a50c262-1a91-4dad-91ec-6530d981415a","Type":"ContainerDied","Data":"259a6976c0342102cdd656ddce3e8b7ae4bcd91eef21326c8f9dac4a2bb40195"} Dec 27 08:15:42 crc kubenswrapper[4934]: I1227 08:15:42.913940 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="259a6976c0342102cdd656ddce3e8b7ae4bcd91eef21326c8f9dac4a2bb40195" Dec 27 08:15:42 crc kubenswrapper[4934]: I1227 08:15:42.913687 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5" Dec 27 08:15:43 crc kubenswrapper[4934]: I1227 08:15:43.032784 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq"] Dec 27 08:15:43 crc kubenswrapper[4934]: E1227 08:15:43.033539 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1457a36-6d7f-4770-bca1-d7bc5919bb4c" containerName="collect-profiles" Dec 27 08:15:43 crc kubenswrapper[4934]: I1227 08:15:43.033572 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1457a36-6d7f-4770-bca1-d7bc5919bb4c" containerName="collect-profiles" Dec 27 08:15:43 crc kubenswrapper[4934]: E1227 08:15:43.033622 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a50c262-1a91-4dad-91ec-6530d981415a" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 27 08:15:43 crc kubenswrapper[4934]: I1227 08:15:43.033636 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a50c262-1a91-4dad-91ec-6530d981415a" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 27 08:15:43 crc kubenswrapper[4934]: I1227 08:15:43.034034 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a50c262-1a91-4dad-91ec-6530d981415a" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 27 08:15:43 crc kubenswrapper[4934]: I1227 08:15:43.034114 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1457a36-6d7f-4770-bca1-d7bc5919bb4c" containerName="collect-profiles" Dec 27 08:15:43 crc kubenswrapper[4934]: I1227 08:15:43.043379 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq" Dec 27 08:15:43 crc kubenswrapper[4934]: I1227 08:15:43.050661 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-98ddn" Dec 27 08:15:43 crc kubenswrapper[4934]: I1227 08:15:43.050987 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 27 08:15:43 crc kubenswrapper[4934]: I1227 08:15:43.051343 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 27 08:15:43 crc kubenswrapper[4934]: I1227 08:15:43.051618 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 27 08:15:43 crc kubenswrapper[4934]: I1227 08:15:43.069227 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq"] Dec 27 08:15:43 crc kubenswrapper[4934]: I1227 08:15:43.190843 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d482978-80ac-442d-817f-70f4270892dd-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq\" (UID: \"3d482978-80ac-442d-817f-70f4270892dd\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq" Dec 27 08:15:43 crc kubenswrapper[4934]: I1227 08:15:43.191042 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d482978-80ac-442d-817f-70f4270892dd-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq\" (UID: \"3d482978-80ac-442d-817f-70f4270892dd\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq" Dec 27 08:15:43 crc kubenswrapper[4934]: I1227 08:15:43.191168 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgsvx\" (UniqueName: \"kubernetes.io/projected/3d482978-80ac-442d-817f-70f4270892dd-kube-api-access-xgsvx\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq\" (UID: \"3d482978-80ac-442d-817f-70f4270892dd\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq" Dec 27 08:15:43 crc kubenswrapper[4934]: I1227 08:15:43.293628 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d482978-80ac-442d-817f-70f4270892dd-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq\" (UID: \"3d482978-80ac-442d-817f-70f4270892dd\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq" Dec 27 08:15:43 crc kubenswrapper[4934]: I1227 08:15:43.293837 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d482978-80ac-442d-817f-70f4270892dd-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq\" (UID: \"3d482978-80ac-442d-817f-70f4270892dd\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq" Dec 27 08:15:43 crc kubenswrapper[4934]: I1227 08:15:43.293925 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgsvx\" (UniqueName: \"kubernetes.io/projected/3d482978-80ac-442d-817f-70f4270892dd-kube-api-access-xgsvx\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq\" (UID: \"3d482978-80ac-442d-817f-70f4270892dd\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq" Dec 27 08:15:43 crc kubenswrapper[4934]: I1227 08:15:43.300871 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d482978-80ac-442d-817f-70f4270892dd-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq\" (UID: \"3d482978-80ac-442d-817f-70f4270892dd\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq" Dec 27 08:15:43 crc kubenswrapper[4934]: I1227 08:15:43.306542 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d482978-80ac-442d-817f-70f4270892dd-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq\" (UID: \"3d482978-80ac-442d-817f-70f4270892dd\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq" Dec 27 08:15:43 crc kubenswrapper[4934]: I1227 08:15:43.315268 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgsvx\" (UniqueName: \"kubernetes.io/projected/3d482978-80ac-442d-817f-70f4270892dd-kube-api-access-xgsvx\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq\" (UID: \"3d482978-80ac-442d-817f-70f4270892dd\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq" Dec 27 08:15:43 crc kubenswrapper[4934]: I1227 08:15:43.365960 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq" Dec 27 08:15:44 crc kubenswrapper[4934]: I1227 08:15:44.008816 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq"] Dec 27 08:15:44 crc kubenswrapper[4934]: I1227 08:15:44.012662 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 27 08:15:44 crc kubenswrapper[4934]: I1227 08:15:44.943298 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq" event={"ID":"3d482978-80ac-442d-817f-70f4270892dd","Type":"ContainerStarted","Data":"8aea898baffba6c6f2df1b64ec29d763c91e3940da4ab250cdfc551c8f7fdad1"} Dec 27 08:15:45 crc kubenswrapper[4934]: I1227 08:15:45.330056 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:15:45 crc kubenswrapper[4934]: I1227 08:15:45.330169 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:15:45 crc kubenswrapper[4934]: I1227 08:15:45.990384 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq" event={"ID":"3d482978-80ac-442d-817f-70f4270892dd","Type":"ContainerStarted","Data":"08a76defdd9a81219ac8d5802d9cf8a84f28d544a20f2689e5e74c7ac7139c3d"} Dec 27 08:15:46 crc kubenswrapper[4934]: I1227 08:15:46.031582 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq" podStartSLOduration=2.363122321 podStartE2EDuration="3.031552235s" podCreationTimestamp="2025-12-27 08:15:43 +0000 UTC" firstStartedPulling="2025-12-27 08:15:44.012345874 +0000 UTC m=+2004.832786478" lastFinishedPulling="2025-12-27 08:15:44.680775798 +0000 UTC m=+2005.501216392" observedRunningTime="2025-12-27 08:15:46.020025208 +0000 UTC m=+2006.840465862" watchObservedRunningTime="2025-12-27 08:15:46.031552235 +0000 UTC m=+2006.851992869" Dec 27 08:15:52 crc kubenswrapper[4934]: I1227 08:15:52.072891 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-zkwng"] Dec 27 08:15:52 crc kubenswrapper[4934]: I1227 08:15:52.089351 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-6kjfr"] Dec 27 08:15:52 crc kubenswrapper[4934]: I1227 08:15:52.104032 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-6kjfr"] Dec 27 08:15:52 crc kubenswrapper[4934]: I1227 08:15:52.113275 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-zkwng"] Dec 27 08:15:52 crc kubenswrapper[4934]: I1227 08:15:52.122177 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-xbmch"] Dec 27 08:15:52 crc kubenswrapper[4934]: I1227 08:15:52.130615 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-xbmch"] Dec 27 08:15:53 crc kubenswrapper[4934]: I1227 08:15:53.484347 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cbaa946-14d7-4839-9541-772b8c5755c0" path="/var/lib/kubelet/pods/5cbaa946-14d7-4839-9541-772b8c5755c0/volumes" Dec 27 08:15:53 crc kubenswrapper[4934]: I1227 08:15:53.487303 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e8b0569-1879-41e6-a91c-7e1b812ff6d5" path="/var/lib/kubelet/pods/8e8b0569-1879-41e6-a91c-7e1b812ff6d5/volumes" Dec 27 08:15:53 crc kubenswrapper[4934]: I1227 08:15:53.488325 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acbab7d1-71c2-4bf6-888b-b0e45005d374" path="/var/lib/kubelet/pods/acbab7d1-71c2-4bf6-888b-b0e45005d374/volumes" Dec 27 08:15:59 crc kubenswrapper[4934]: I1227 08:15:59.045027 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-5b72-account-create-xsfb7"] Dec 27 08:15:59 crc kubenswrapper[4934]: I1227 08:15:59.058537 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-084f-account-create-9l2gc"] Dec 27 08:15:59 crc kubenswrapper[4934]: I1227 08:15:59.069180 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-5b72-account-create-xsfb7"] Dec 27 08:15:59 crc kubenswrapper[4934]: I1227 08:15:59.079014 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-084f-account-create-9l2gc"] Dec 27 08:15:59 crc kubenswrapper[4934]: I1227 08:15:59.088320 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-5cdf-account-create-hdhg5"] Dec 27 08:15:59 crc kubenswrapper[4934]: I1227 08:15:59.097332 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-5cdf-account-create-hdhg5"] Dec 27 08:15:59 crc kubenswrapper[4934]: I1227 08:15:59.505125 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c9fcfdd-943b-42b6-9741-74feaf6d90f7" path="/var/lib/kubelet/pods/0c9fcfdd-943b-42b6-9741-74feaf6d90f7/volumes" Dec 27 08:15:59 crc kubenswrapper[4934]: I1227 08:15:59.509257 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42a96968-f7ee-40bf-9d60-b662e888db0d" path="/var/lib/kubelet/pods/42a96968-f7ee-40bf-9d60-b662e888db0d/volumes" Dec 27 08:15:59 crc kubenswrapper[4934]: I1227 08:15:59.510847 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74a539ac-109e-4442-ac71-209a2ed6a27d" path="/var/lib/kubelet/pods/74a539ac-109e-4442-ac71-209a2ed6a27d/volumes" Dec 27 08:16:15 crc kubenswrapper[4934]: I1227 08:16:15.330668 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:16:15 crc kubenswrapper[4934]: I1227 08:16:15.331246 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:16:25 crc kubenswrapper[4934]: I1227 08:16:25.076280 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cm8ph"] Dec 27 08:16:25 crc kubenswrapper[4934]: I1227 08:16:25.090003 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-cm8ph"] Dec 27 08:16:25 crc kubenswrapper[4934]: I1227 08:16:25.495335 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85e47792-15be-4b6b-8a8e-8c05c4ae4605" path="/var/lib/kubelet/pods/85e47792-15be-4b6b-8a8e-8c05c4ae4605/volumes" Dec 27 08:16:27 crc kubenswrapper[4934]: I1227 08:16:27.076013 4934 scope.go:117] "RemoveContainer" containerID="9559761b33b407ededc4d3d58c8f1366c61c5d156b946dc7bafd3405ad9b4c9b" Dec 27 08:16:27 crc kubenswrapper[4934]: I1227 08:16:27.130026 4934 scope.go:117] "RemoveContainer" containerID="a25540b169ccd286d2daa333158ae981f01dc148e21953b9855b096d30040eb1" Dec 27 08:16:27 crc kubenswrapper[4934]: I1227 08:16:27.223455 4934 scope.go:117] "RemoveContainer" containerID="0d4b1755b557ba1c420b5022b2131c77712aced9ec65c3d60cb4378ebcc83d68" Dec 27 08:16:27 crc kubenswrapper[4934]: I1227 08:16:27.258446 4934 scope.go:117] "RemoveContainer" containerID="5dadcba05a120b34a236db90c5a70fb53fbc446fd5c98839d70ace484e18f906" Dec 27 08:16:27 crc kubenswrapper[4934]: I1227 08:16:27.315725 4934 scope.go:117] "RemoveContainer" containerID="851f13e7d7fd0da526e524752d71b5714f8fdde4d26e110e7f6a43c94d24d2f5" Dec 27 08:16:27 crc kubenswrapper[4934]: I1227 08:16:27.365011 4934 scope.go:117] "RemoveContainer" containerID="9e228948ddcf016bf0a216af729f9f9a1f35ee7a0eb7def58d6a6cff43b65cd3" Dec 27 08:16:27 crc kubenswrapper[4934]: I1227 08:16:27.416621 4934 scope.go:117] "RemoveContainer" containerID="198266d3b59d65fb02b064c61f2f6e47c2b54fd06aa20323c6b09a9b554e7050" Dec 27 08:16:38 crc kubenswrapper[4934]: I1227 08:16:38.055726 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-4vl7v"] Dec 27 08:16:38 crc kubenswrapper[4934]: I1227 08:16:38.067673 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-4vl7v"] Dec 27 08:16:39 crc kubenswrapper[4934]: I1227 08:16:39.493416 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38941274-cf93-4985-9e25-f9aad3dcb394" path="/var/lib/kubelet/pods/38941274-cf93-4985-9e25-f9aad3dcb394/volumes" Dec 27 08:16:45 crc kubenswrapper[4934]: I1227 08:16:45.330348 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:16:45 crc kubenswrapper[4934]: I1227 08:16:45.331077 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:16:45 crc kubenswrapper[4934]: I1227 08:16:45.331192 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 08:16:45 crc kubenswrapper[4934]: I1227 08:16:45.332516 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"deb6fb2d2d95a934db9a65eebdf5e58dc64e13629a44e74765a4e64756d050bb"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 08:16:45 crc kubenswrapper[4934]: I1227 08:16:45.332618 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://deb6fb2d2d95a934db9a65eebdf5e58dc64e13629a44e74765a4e64756d050bb" gracePeriod=600 Dec 27 08:16:45 crc kubenswrapper[4934]: I1227 08:16:45.878858 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="deb6fb2d2d95a934db9a65eebdf5e58dc64e13629a44e74765a4e64756d050bb" exitCode=0 Dec 27 08:16:45 crc kubenswrapper[4934]: I1227 08:16:45.879269 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"deb6fb2d2d95a934db9a65eebdf5e58dc64e13629a44e74765a4e64756d050bb"} Dec 27 08:16:45 crc kubenswrapper[4934]: I1227 08:16:45.879604 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6"} Dec 27 08:16:45 crc kubenswrapper[4934]: I1227 08:16:45.879653 4934 scope.go:117] "RemoveContainer" containerID="816dcfc94a2634071db4437201d7c2bbac52fd17f029aa5ef7764f16e829b7ab" Dec 27 08:16:47 crc kubenswrapper[4934]: I1227 08:16:47.044258 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-2566-account-create-rtm45"] Dec 27 08:16:47 crc kubenswrapper[4934]: I1227 08:16:47.057950 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-2566-account-create-rtm45"] Dec 27 08:16:47 crc kubenswrapper[4934]: I1227 08:16:47.490594 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72321394-c1c2-4254-9e2a-10617a3ba8c4" path="/var/lib/kubelet/pods/72321394-c1c2-4254-9e2a-10617a3ba8c4/volumes" Dec 27 08:16:48 crc kubenswrapper[4934]: I1227 08:16:48.042147 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6kzb4"] Dec 27 08:16:48 crc kubenswrapper[4934]: I1227 08:16:48.054765 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-6kzb4"] Dec 27 08:16:49 crc kubenswrapper[4934]: I1227 08:16:49.042136 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-qxkzj"] Dec 27 08:16:49 crc kubenswrapper[4934]: I1227 08:16:49.053027 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-qxkzj"] Dec 27 08:16:49 crc kubenswrapper[4934]: I1227 08:16:49.487320 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52a3f24b-77b1-467e-b9ce-03e2ae5fe53e" path="/var/lib/kubelet/pods/52a3f24b-77b1-467e-b9ce-03e2ae5fe53e/volumes" Dec 27 08:16:49 crc kubenswrapper[4934]: I1227 08:16:49.488197 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94199ff4-bea2-4054-9752-7f9e5cd73d99" path="/var/lib/kubelet/pods/94199ff4-bea2-4054-9752-7f9e5cd73d99/volumes" Dec 27 08:17:03 crc kubenswrapper[4934]: I1227 08:17:03.156798 4934 generic.go:334] "Generic (PLEG): container finished" podID="3d482978-80ac-442d-817f-70f4270892dd" containerID="08a76defdd9a81219ac8d5802d9cf8a84f28d544a20f2689e5e74c7ac7139c3d" exitCode=0 Dec 27 08:17:03 crc kubenswrapper[4934]: I1227 08:17:03.156936 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq" event={"ID":"3d482978-80ac-442d-817f-70f4270892dd","Type":"ContainerDied","Data":"08a76defdd9a81219ac8d5802d9cf8a84f28d544a20f2689e5e74c7ac7139c3d"} Dec 27 08:17:04 crc kubenswrapper[4934]: I1227 08:17:04.713948 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq" Dec 27 08:17:04 crc kubenswrapper[4934]: I1227 08:17:04.858368 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d482978-80ac-442d-817f-70f4270892dd-inventory\") pod \"3d482978-80ac-442d-817f-70f4270892dd\" (UID: \"3d482978-80ac-442d-817f-70f4270892dd\") " Dec 27 08:17:04 crc kubenswrapper[4934]: I1227 08:17:04.858836 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgsvx\" (UniqueName: \"kubernetes.io/projected/3d482978-80ac-442d-817f-70f4270892dd-kube-api-access-xgsvx\") pod \"3d482978-80ac-442d-817f-70f4270892dd\" (UID: \"3d482978-80ac-442d-817f-70f4270892dd\") " Dec 27 08:17:04 crc kubenswrapper[4934]: I1227 08:17:04.859279 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d482978-80ac-442d-817f-70f4270892dd-ssh-key\") pod \"3d482978-80ac-442d-817f-70f4270892dd\" (UID: \"3d482978-80ac-442d-817f-70f4270892dd\") " Dec 27 08:17:04 crc kubenswrapper[4934]: I1227 08:17:04.870366 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d482978-80ac-442d-817f-70f4270892dd-kube-api-access-xgsvx" (OuterVolumeSpecName: "kube-api-access-xgsvx") pod "3d482978-80ac-442d-817f-70f4270892dd" (UID: "3d482978-80ac-442d-817f-70f4270892dd"). InnerVolumeSpecName "kube-api-access-xgsvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:17:04 crc kubenswrapper[4934]: I1227 08:17:04.892891 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d482978-80ac-442d-817f-70f4270892dd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3d482978-80ac-442d-817f-70f4270892dd" (UID: "3d482978-80ac-442d-817f-70f4270892dd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:17:04 crc kubenswrapper[4934]: I1227 08:17:04.929043 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d482978-80ac-442d-817f-70f4270892dd-inventory" (OuterVolumeSpecName: "inventory") pod "3d482978-80ac-442d-817f-70f4270892dd" (UID: "3d482978-80ac-442d-817f-70f4270892dd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:17:04 crc kubenswrapper[4934]: I1227 08:17:04.963194 4934 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d482978-80ac-442d-817f-70f4270892dd-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 27 08:17:04 crc kubenswrapper[4934]: I1227 08:17:04.963233 4934 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d482978-80ac-442d-817f-70f4270892dd-inventory\") on node \"crc\" DevicePath \"\"" Dec 27 08:17:04 crc kubenswrapper[4934]: I1227 08:17:04.963249 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgsvx\" (UniqueName: \"kubernetes.io/projected/3d482978-80ac-442d-817f-70f4270892dd-kube-api-access-xgsvx\") on node \"crc\" DevicePath \"\"" Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.188263 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq" event={"ID":"3d482978-80ac-442d-817f-70f4270892dd","Type":"ContainerDied","Data":"8aea898baffba6c6f2df1b64ec29d763c91e3940da4ab250cdfc551c8f7fdad1"} Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.188302 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8aea898baffba6c6f2df1b64ec29d763c91e3940da4ab250cdfc551c8f7fdad1" Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.188372 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq" Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.310890 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz"] Dec 27 08:17:05 crc kubenswrapper[4934]: E1227 08:17:05.311465 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d482978-80ac-442d-817f-70f4270892dd" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.311487 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d482978-80ac-442d-817f-70f4270892dd" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.311749 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d482978-80ac-442d-817f-70f4270892dd" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.312706 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz" Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.315160 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.315230 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.315290 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.317291 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-98ddn" Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.334533 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz"] Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.374900 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6174eb59-8d05-409e-a838-807ef965ee5c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz\" (UID: \"6174eb59-8d05-409e-a838-807ef965ee5c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz" Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.374991 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6174eb59-8d05-409e-a838-807ef965ee5c-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz\" (UID: \"6174eb59-8d05-409e-a838-807ef965ee5c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz" Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.375371 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hclp\" (UniqueName: \"kubernetes.io/projected/6174eb59-8d05-409e-a838-807ef965ee5c-kube-api-access-5hclp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz\" (UID: \"6174eb59-8d05-409e-a838-807ef965ee5c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz" Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.477211 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6174eb59-8d05-409e-a838-807ef965ee5c-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz\" (UID: \"6174eb59-8d05-409e-a838-807ef965ee5c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz" Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.477706 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hclp\" (UniqueName: \"kubernetes.io/projected/6174eb59-8d05-409e-a838-807ef965ee5c-kube-api-access-5hclp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz\" (UID: \"6174eb59-8d05-409e-a838-807ef965ee5c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz" Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.477794 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6174eb59-8d05-409e-a838-807ef965ee5c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz\" (UID: \"6174eb59-8d05-409e-a838-807ef965ee5c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz" Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.483018 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6174eb59-8d05-409e-a838-807ef965ee5c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz\" (UID: \"6174eb59-8d05-409e-a838-807ef965ee5c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz" Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.483037 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6174eb59-8d05-409e-a838-807ef965ee5c-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz\" (UID: \"6174eb59-8d05-409e-a838-807ef965ee5c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz" Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.499231 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hclp\" (UniqueName: \"kubernetes.io/projected/6174eb59-8d05-409e-a838-807ef965ee5c-kube-api-access-5hclp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz\" (UID: \"6174eb59-8d05-409e-a838-807ef965ee5c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz" Dec 27 08:17:05 crc kubenswrapper[4934]: I1227 08:17:05.644299 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz" Dec 27 08:17:06 crc kubenswrapper[4934]: I1227 08:17:06.370186 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz"] Dec 27 08:17:07 crc kubenswrapper[4934]: I1227 08:17:07.218177 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz" event={"ID":"6174eb59-8d05-409e-a838-807ef965ee5c","Type":"ContainerStarted","Data":"17caca0940be0b78b4af75d879b5d82d2f3e5981d68fbc719b7a9b543d206c35"} Dec 27 08:17:07 crc kubenswrapper[4934]: I1227 08:17:07.218610 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz" event={"ID":"6174eb59-8d05-409e-a838-807ef965ee5c","Type":"ContainerStarted","Data":"d26bba1d71706ac4ac2ffe18237cd8b0fe6c315baf52b62e3643754be995c69e"} Dec 27 08:17:13 crc kubenswrapper[4934]: I1227 08:17:13.309154 4934 generic.go:334] "Generic (PLEG): container finished" podID="6174eb59-8d05-409e-a838-807ef965ee5c" containerID="17caca0940be0b78b4af75d879b5d82d2f3e5981d68fbc719b7a9b543d206c35" exitCode=0 Dec 27 08:17:13 crc kubenswrapper[4934]: I1227 08:17:13.309233 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz" event={"ID":"6174eb59-8d05-409e-a838-807ef965ee5c","Type":"ContainerDied","Data":"17caca0940be0b78b4af75d879b5d82d2f3e5981d68fbc719b7a9b543d206c35"} Dec 27 08:17:14 crc kubenswrapper[4934]: I1227 08:17:14.940237 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz" Dec 27 08:17:14 crc kubenswrapper[4934]: I1227 08:17:14.967403 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6174eb59-8d05-409e-a838-807ef965ee5c-ssh-key\") pod \"6174eb59-8d05-409e-a838-807ef965ee5c\" (UID: \"6174eb59-8d05-409e-a838-807ef965ee5c\") " Dec 27 08:17:14 crc kubenswrapper[4934]: I1227 08:17:14.967499 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6174eb59-8d05-409e-a838-807ef965ee5c-inventory\") pod \"6174eb59-8d05-409e-a838-807ef965ee5c\" (UID: \"6174eb59-8d05-409e-a838-807ef965ee5c\") " Dec 27 08:17:14 crc kubenswrapper[4934]: I1227 08:17:14.967621 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hclp\" (UniqueName: \"kubernetes.io/projected/6174eb59-8d05-409e-a838-807ef965ee5c-kube-api-access-5hclp\") pod \"6174eb59-8d05-409e-a838-807ef965ee5c\" (UID: \"6174eb59-8d05-409e-a838-807ef965ee5c\") " Dec 27 08:17:14 crc kubenswrapper[4934]: I1227 08:17:14.980457 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6174eb59-8d05-409e-a838-807ef965ee5c-kube-api-access-5hclp" (OuterVolumeSpecName: "kube-api-access-5hclp") pod "6174eb59-8d05-409e-a838-807ef965ee5c" (UID: "6174eb59-8d05-409e-a838-807ef965ee5c"). InnerVolumeSpecName "kube-api-access-5hclp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.028404 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6174eb59-8d05-409e-a838-807ef965ee5c-inventory" (OuterVolumeSpecName: "inventory") pod "6174eb59-8d05-409e-a838-807ef965ee5c" (UID: "6174eb59-8d05-409e-a838-807ef965ee5c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.028627 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6174eb59-8d05-409e-a838-807ef965ee5c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6174eb59-8d05-409e-a838-807ef965ee5c" (UID: "6174eb59-8d05-409e-a838-807ef965ee5c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.071328 4934 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6174eb59-8d05-409e-a838-807ef965ee5c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.071383 4934 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6174eb59-8d05-409e-a838-807ef965ee5c-inventory\") on node \"crc\" DevicePath \"\"" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.071401 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hclp\" (UniqueName: \"kubernetes.io/projected/6174eb59-8d05-409e-a838-807ef965ee5c-kube-api-access-5hclp\") on node \"crc\" DevicePath \"\"" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.337390 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz" event={"ID":"6174eb59-8d05-409e-a838-807ef965ee5c","Type":"ContainerDied","Data":"d26bba1d71706ac4ac2ffe18237cd8b0fe6c315baf52b62e3643754be995c69e"} Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.337444 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d26bba1d71706ac4ac2ffe18237cd8b0fe6c315baf52b62e3643754be995c69e" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.338001 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.436918 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-7cw86"] Dec 27 08:17:15 crc kubenswrapper[4934]: E1227 08:17:15.437818 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6174eb59-8d05-409e-a838-807ef965ee5c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.437839 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="6174eb59-8d05-409e-a838-807ef965ee5c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.438183 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="6174eb59-8d05-409e-a838-807ef965ee5c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.439375 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7cw86" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.446568 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.446652 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-98ddn" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.446961 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.447425 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.481279 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-7cw86"] Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.481345 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41797816-91e0-4a3e-ad30-91da21255b91-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7cw86\" (UID: \"41797816-91e0-4a3e-ad30-91da21255b91\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7cw86" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.481505 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zr76\" (UniqueName: \"kubernetes.io/projected/41797816-91e0-4a3e-ad30-91da21255b91-kube-api-access-9zr76\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7cw86\" (UID: \"41797816-91e0-4a3e-ad30-91da21255b91\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7cw86" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.481705 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41797816-91e0-4a3e-ad30-91da21255b91-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7cw86\" (UID: \"41797816-91e0-4a3e-ad30-91da21255b91\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7cw86" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.584225 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zr76\" (UniqueName: \"kubernetes.io/projected/41797816-91e0-4a3e-ad30-91da21255b91-kube-api-access-9zr76\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7cw86\" (UID: \"41797816-91e0-4a3e-ad30-91da21255b91\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7cw86" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.584344 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41797816-91e0-4a3e-ad30-91da21255b91-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7cw86\" (UID: \"41797816-91e0-4a3e-ad30-91da21255b91\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7cw86" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.584472 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41797816-91e0-4a3e-ad30-91da21255b91-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7cw86\" (UID: \"41797816-91e0-4a3e-ad30-91da21255b91\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7cw86" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.590104 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41797816-91e0-4a3e-ad30-91da21255b91-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7cw86\" (UID: \"41797816-91e0-4a3e-ad30-91da21255b91\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7cw86" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.602704 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41797816-91e0-4a3e-ad30-91da21255b91-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7cw86\" (UID: \"41797816-91e0-4a3e-ad30-91da21255b91\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7cw86" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.604400 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zr76\" (UniqueName: \"kubernetes.io/projected/41797816-91e0-4a3e-ad30-91da21255b91-kube-api-access-9zr76\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7cw86\" (UID: \"41797816-91e0-4a3e-ad30-91da21255b91\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7cw86" Dec 27 08:17:15 crc kubenswrapper[4934]: I1227 08:17:15.765805 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7cw86" Dec 27 08:17:16 crc kubenswrapper[4934]: I1227 08:17:16.377061 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-7cw86"] Dec 27 08:17:16 crc kubenswrapper[4934]: W1227 08:17:16.386419 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41797816_91e0_4a3e_ad30_91da21255b91.slice/crio-beaeafa13f316194f95dc9863bc98f59d22280ba751c1a1bb38a6e22d6fdc686 WatchSource:0}: Error finding container beaeafa13f316194f95dc9863bc98f59d22280ba751c1a1bb38a6e22d6fdc686: Status 404 returned error can't find the container with id beaeafa13f316194f95dc9863bc98f59d22280ba751c1a1bb38a6e22d6fdc686 Dec 27 08:17:17 crc kubenswrapper[4934]: I1227 08:17:17.372949 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7cw86" event={"ID":"41797816-91e0-4a3e-ad30-91da21255b91","Type":"ContainerStarted","Data":"74a9c1fb4e03b34493c06362524cd0f14200cb047798854fa5eed586c526204c"} Dec 27 08:17:17 crc kubenswrapper[4934]: I1227 08:17:17.373663 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7cw86" event={"ID":"41797816-91e0-4a3e-ad30-91da21255b91","Type":"ContainerStarted","Data":"beaeafa13f316194f95dc9863bc98f59d22280ba751c1a1bb38a6e22d6fdc686"} Dec 27 08:17:17 crc kubenswrapper[4934]: I1227 08:17:17.398037 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7cw86" podStartSLOduration=1.966745704 podStartE2EDuration="2.398005494s" podCreationTimestamp="2025-12-27 08:17:15 +0000 UTC" firstStartedPulling="2025-12-27 08:17:16.400969774 +0000 UTC m=+2097.221410368" lastFinishedPulling="2025-12-27 08:17:16.832229554 +0000 UTC m=+2097.652670158" observedRunningTime="2025-12-27 08:17:17.393572573 +0000 UTC m=+2098.214013207" watchObservedRunningTime="2025-12-27 08:17:17.398005494 +0000 UTC m=+2098.218446118" Dec 27 08:17:27 crc kubenswrapper[4934]: I1227 08:17:27.662003 4934 scope.go:117] "RemoveContainer" containerID="b2b5df90fe5da94b89d952defbf811bc55a7d5d080faf0300c01ddbb0551a5d1" Dec 27 08:17:27 crc kubenswrapper[4934]: I1227 08:17:27.702005 4934 scope.go:117] "RemoveContainer" containerID="201c806d5834ec84e492faab44bdf7372b91b98e98442e86462b31551fb92d58" Dec 27 08:17:27 crc kubenswrapper[4934]: I1227 08:17:27.788720 4934 scope.go:117] "RemoveContainer" containerID="efdee1bf217bc262bb57f92fa39bb1ff916bdd3ee785824c29c4a087d23c6b74" Dec 27 08:17:27 crc kubenswrapper[4934]: I1227 08:17:27.835393 4934 scope.go:117] "RemoveContainer" containerID="a3ba409c30638f31737d02cdbbb7f3b230286a84cb70201714fe63efda870e2c" Dec 27 08:17:32 crc kubenswrapper[4934]: I1227 08:17:32.059804 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-plls4"] Dec 27 08:17:32 crc kubenswrapper[4934]: I1227 08:17:32.075768 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-plls4"] Dec 27 08:17:33 crc kubenswrapper[4934]: I1227 08:17:33.488998 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e" path="/var/lib/kubelet/pods/1b85e56c-313b-4bac-b1d8-bd5d7c23ba6e/volumes" Dec 27 08:17:54 crc kubenswrapper[4934]: I1227 08:17:54.912850 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hhp5b"] Dec 27 08:17:54 crc kubenswrapper[4934]: I1227 08:17:54.917921 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hhp5b" Dec 27 08:17:54 crc kubenswrapper[4934]: I1227 08:17:54.923369 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hhp5b"] Dec 27 08:17:55 crc kubenswrapper[4934]: I1227 08:17:55.035899 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a018a29-f75d-47b4-9af4-e170e9110551-utilities\") pod \"certified-operators-hhp5b\" (UID: \"8a018a29-f75d-47b4-9af4-e170e9110551\") " pod="openshift-marketplace/certified-operators-hhp5b" Dec 27 08:17:55 crc kubenswrapper[4934]: I1227 08:17:55.036066 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ff27\" (UniqueName: \"kubernetes.io/projected/8a018a29-f75d-47b4-9af4-e170e9110551-kube-api-access-5ff27\") pod \"certified-operators-hhp5b\" (UID: \"8a018a29-f75d-47b4-9af4-e170e9110551\") " pod="openshift-marketplace/certified-operators-hhp5b" Dec 27 08:17:55 crc kubenswrapper[4934]: I1227 08:17:55.036127 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a018a29-f75d-47b4-9af4-e170e9110551-catalog-content\") pod \"certified-operators-hhp5b\" (UID: \"8a018a29-f75d-47b4-9af4-e170e9110551\") " pod="openshift-marketplace/certified-operators-hhp5b" Dec 27 08:17:55 crc kubenswrapper[4934]: I1227 08:17:55.138284 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a018a29-f75d-47b4-9af4-e170e9110551-utilities\") pod \"certified-operators-hhp5b\" (UID: \"8a018a29-f75d-47b4-9af4-e170e9110551\") " pod="openshift-marketplace/certified-operators-hhp5b" Dec 27 08:17:55 crc kubenswrapper[4934]: I1227 08:17:55.138673 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ff27\" (UniqueName: \"kubernetes.io/projected/8a018a29-f75d-47b4-9af4-e170e9110551-kube-api-access-5ff27\") pod \"certified-operators-hhp5b\" (UID: \"8a018a29-f75d-47b4-9af4-e170e9110551\") " pod="openshift-marketplace/certified-operators-hhp5b" Dec 27 08:17:55 crc kubenswrapper[4934]: I1227 08:17:55.138815 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a018a29-f75d-47b4-9af4-e170e9110551-catalog-content\") pod \"certified-operators-hhp5b\" (UID: \"8a018a29-f75d-47b4-9af4-e170e9110551\") " pod="openshift-marketplace/certified-operators-hhp5b" Dec 27 08:17:55 crc kubenswrapper[4934]: I1227 08:17:55.138886 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a018a29-f75d-47b4-9af4-e170e9110551-utilities\") pod \"certified-operators-hhp5b\" (UID: \"8a018a29-f75d-47b4-9af4-e170e9110551\") " pod="openshift-marketplace/certified-operators-hhp5b" Dec 27 08:17:55 crc kubenswrapper[4934]: I1227 08:17:55.139397 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a018a29-f75d-47b4-9af4-e170e9110551-catalog-content\") pod \"certified-operators-hhp5b\" (UID: \"8a018a29-f75d-47b4-9af4-e170e9110551\") " pod="openshift-marketplace/certified-operators-hhp5b" Dec 27 08:17:55 crc kubenswrapper[4934]: I1227 08:17:55.159104 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ff27\" (UniqueName: \"kubernetes.io/projected/8a018a29-f75d-47b4-9af4-e170e9110551-kube-api-access-5ff27\") pod \"certified-operators-hhp5b\" (UID: \"8a018a29-f75d-47b4-9af4-e170e9110551\") " pod="openshift-marketplace/certified-operators-hhp5b" Dec 27 08:17:55 crc kubenswrapper[4934]: I1227 08:17:55.251532 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hhp5b" Dec 27 08:17:55 crc kubenswrapper[4934]: I1227 08:17:55.827291 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hhp5b"] Dec 27 08:17:55 crc kubenswrapper[4934]: I1227 08:17:55.870175 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hhp5b" event={"ID":"8a018a29-f75d-47b4-9af4-e170e9110551","Type":"ContainerStarted","Data":"6e6a1513b42af1b724c7a951def4877b11c251157bdcc688ff96ab305b5ece9d"} Dec 27 08:17:56 crc kubenswrapper[4934]: I1227 08:17:56.880753 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hhp5b" event={"ID":"8a018a29-f75d-47b4-9af4-e170e9110551","Type":"ContainerDied","Data":"aec809f3a76f8f1291d6c7d892ad88fd4f7d8ab9242f76449ead528917ad28f0"} Dec 27 08:17:56 crc kubenswrapper[4934]: I1227 08:17:56.880579 4934 generic.go:334] "Generic (PLEG): container finished" podID="8a018a29-f75d-47b4-9af4-e170e9110551" containerID="aec809f3a76f8f1291d6c7d892ad88fd4f7d8ab9242f76449ead528917ad28f0" exitCode=0 Dec 27 08:17:58 crc kubenswrapper[4934]: I1227 08:17:58.900940 4934 generic.go:334] "Generic (PLEG): container finished" podID="8a018a29-f75d-47b4-9af4-e170e9110551" containerID="ab7022b7090c09020243fecf51234bb3388f36232d6f18cca805a0fda9fc515e" exitCode=0 Dec 27 08:17:58 crc kubenswrapper[4934]: I1227 08:17:58.901532 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hhp5b" event={"ID":"8a018a29-f75d-47b4-9af4-e170e9110551","Type":"ContainerDied","Data":"ab7022b7090c09020243fecf51234bb3388f36232d6f18cca805a0fda9fc515e"} Dec 27 08:17:59 crc kubenswrapper[4934]: I1227 08:17:59.916234 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hhp5b" event={"ID":"8a018a29-f75d-47b4-9af4-e170e9110551","Type":"ContainerStarted","Data":"75dad1f790c0bb0acb68803def294aa97efe47b8cb06d9deba69c5f39d54028e"} Dec 27 08:17:59 crc kubenswrapper[4934]: I1227 08:17:59.919190 4934 generic.go:334] "Generic (PLEG): container finished" podID="41797816-91e0-4a3e-ad30-91da21255b91" containerID="74a9c1fb4e03b34493c06362524cd0f14200cb047798854fa5eed586c526204c" exitCode=0 Dec 27 08:17:59 crc kubenswrapper[4934]: I1227 08:17:59.919277 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7cw86" event={"ID":"41797816-91e0-4a3e-ad30-91da21255b91","Type":"ContainerDied","Data":"74a9c1fb4e03b34493c06362524cd0f14200cb047798854fa5eed586c526204c"} Dec 27 08:17:59 crc kubenswrapper[4934]: I1227 08:17:59.950054 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hhp5b" podStartSLOduration=3.496102766 podStartE2EDuration="5.950024198s" podCreationTimestamp="2025-12-27 08:17:54 +0000 UTC" firstStartedPulling="2025-12-27 08:17:56.882575245 +0000 UTC m=+2137.703015839" lastFinishedPulling="2025-12-27 08:17:59.336496647 +0000 UTC m=+2140.156937271" observedRunningTime="2025-12-27 08:17:59.933869385 +0000 UTC m=+2140.754309979" watchObservedRunningTime="2025-12-27 08:17:59.950024198 +0000 UTC m=+2140.770464802" Dec 27 08:18:01 crc kubenswrapper[4934]: I1227 08:18:01.650675 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7cw86" Dec 27 08:18:01 crc kubenswrapper[4934]: I1227 08:18:01.813847 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41797816-91e0-4a3e-ad30-91da21255b91-ssh-key\") pod \"41797816-91e0-4a3e-ad30-91da21255b91\" (UID: \"41797816-91e0-4a3e-ad30-91da21255b91\") " Dec 27 08:18:01 crc kubenswrapper[4934]: I1227 08:18:01.814073 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zr76\" (UniqueName: \"kubernetes.io/projected/41797816-91e0-4a3e-ad30-91da21255b91-kube-api-access-9zr76\") pod \"41797816-91e0-4a3e-ad30-91da21255b91\" (UID: \"41797816-91e0-4a3e-ad30-91da21255b91\") " Dec 27 08:18:01 crc kubenswrapper[4934]: I1227 08:18:01.814331 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41797816-91e0-4a3e-ad30-91da21255b91-inventory\") pod \"41797816-91e0-4a3e-ad30-91da21255b91\" (UID: \"41797816-91e0-4a3e-ad30-91da21255b91\") " Dec 27 08:18:01 crc kubenswrapper[4934]: I1227 08:18:01.820936 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41797816-91e0-4a3e-ad30-91da21255b91-kube-api-access-9zr76" (OuterVolumeSpecName: "kube-api-access-9zr76") pod "41797816-91e0-4a3e-ad30-91da21255b91" (UID: "41797816-91e0-4a3e-ad30-91da21255b91"). InnerVolumeSpecName "kube-api-access-9zr76". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:18:01 crc kubenswrapper[4934]: I1227 08:18:01.860437 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41797816-91e0-4a3e-ad30-91da21255b91-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "41797816-91e0-4a3e-ad30-91da21255b91" (UID: "41797816-91e0-4a3e-ad30-91da21255b91"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:18:01 crc kubenswrapper[4934]: I1227 08:18:01.885099 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41797816-91e0-4a3e-ad30-91da21255b91-inventory" (OuterVolumeSpecName: "inventory") pod "41797816-91e0-4a3e-ad30-91da21255b91" (UID: "41797816-91e0-4a3e-ad30-91da21255b91"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:18:01 crc kubenswrapper[4934]: I1227 08:18:01.918834 4934 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41797816-91e0-4a3e-ad30-91da21255b91-inventory\") on node \"crc\" DevicePath \"\"" Dec 27 08:18:01 crc kubenswrapper[4934]: I1227 08:18:01.918866 4934 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41797816-91e0-4a3e-ad30-91da21255b91-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 27 08:18:01 crc kubenswrapper[4934]: I1227 08:18:01.918876 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zr76\" (UniqueName: \"kubernetes.io/projected/41797816-91e0-4a3e-ad30-91da21255b91-kube-api-access-9zr76\") on node \"crc\" DevicePath \"\"" Dec 27 08:18:01 crc kubenswrapper[4934]: I1227 08:18:01.942993 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7cw86" event={"ID":"41797816-91e0-4a3e-ad30-91da21255b91","Type":"ContainerDied","Data":"beaeafa13f316194f95dc9863bc98f59d22280ba751c1a1bb38a6e22d6fdc686"} Dec 27 08:18:01 crc kubenswrapper[4934]: I1227 08:18:01.943042 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="beaeafa13f316194f95dc9863bc98f59d22280ba751c1a1bb38a6e22d6fdc686" Dec 27 08:18:01 crc kubenswrapper[4934]: I1227 08:18:01.943152 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7cw86" Dec 27 08:18:02 crc kubenswrapper[4934]: I1227 08:18:02.067826 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d"] Dec 27 08:18:02 crc kubenswrapper[4934]: E1227 08:18:02.068336 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41797816-91e0-4a3e-ad30-91da21255b91" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 27 08:18:02 crc kubenswrapper[4934]: I1227 08:18:02.068353 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="41797816-91e0-4a3e-ad30-91da21255b91" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 27 08:18:02 crc kubenswrapper[4934]: I1227 08:18:02.068651 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="41797816-91e0-4a3e-ad30-91da21255b91" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 27 08:18:02 crc kubenswrapper[4934]: I1227 08:18:02.069505 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d" Dec 27 08:18:02 crc kubenswrapper[4934]: I1227 08:18:02.071873 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-98ddn" Dec 27 08:18:02 crc kubenswrapper[4934]: I1227 08:18:02.072159 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 27 08:18:02 crc kubenswrapper[4934]: I1227 08:18:02.072319 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 27 08:18:02 crc kubenswrapper[4934]: I1227 08:18:02.072858 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 27 08:18:02 crc kubenswrapper[4934]: I1227 08:18:02.086990 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d"] Dec 27 08:18:02 crc kubenswrapper[4934]: I1227 08:18:02.224724 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8af50e00-8101-4c06-b3bc-93fd48753773-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d\" (UID: \"8af50e00-8101-4c06-b3bc-93fd48753773\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d" Dec 27 08:18:02 crc kubenswrapper[4934]: I1227 08:18:02.224960 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8af50e00-8101-4c06-b3bc-93fd48753773-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d\" (UID: \"8af50e00-8101-4c06-b3bc-93fd48753773\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d" Dec 27 08:18:02 crc kubenswrapper[4934]: I1227 08:18:02.225032 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd8ft\" (UniqueName: \"kubernetes.io/projected/8af50e00-8101-4c06-b3bc-93fd48753773-kube-api-access-jd8ft\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d\" (UID: \"8af50e00-8101-4c06-b3bc-93fd48753773\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d" Dec 27 08:18:02 crc kubenswrapper[4934]: I1227 08:18:02.327598 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8af50e00-8101-4c06-b3bc-93fd48753773-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d\" (UID: \"8af50e00-8101-4c06-b3bc-93fd48753773\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d" Dec 27 08:18:02 crc kubenswrapper[4934]: I1227 08:18:02.327657 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd8ft\" (UniqueName: \"kubernetes.io/projected/8af50e00-8101-4c06-b3bc-93fd48753773-kube-api-access-jd8ft\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d\" (UID: \"8af50e00-8101-4c06-b3bc-93fd48753773\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d" Dec 27 08:18:02 crc kubenswrapper[4934]: I1227 08:18:02.327805 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8af50e00-8101-4c06-b3bc-93fd48753773-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d\" (UID: \"8af50e00-8101-4c06-b3bc-93fd48753773\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d" Dec 27 08:18:02 crc kubenswrapper[4934]: I1227 08:18:02.332335 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8af50e00-8101-4c06-b3bc-93fd48753773-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d\" (UID: \"8af50e00-8101-4c06-b3bc-93fd48753773\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d" Dec 27 08:18:02 crc kubenswrapper[4934]: I1227 08:18:02.333268 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8af50e00-8101-4c06-b3bc-93fd48753773-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d\" (UID: \"8af50e00-8101-4c06-b3bc-93fd48753773\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d" Dec 27 08:18:02 crc kubenswrapper[4934]: I1227 08:18:02.362204 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd8ft\" (UniqueName: \"kubernetes.io/projected/8af50e00-8101-4c06-b3bc-93fd48753773-kube-api-access-jd8ft\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d\" (UID: \"8af50e00-8101-4c06-b3bc-93fd48753773\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d" Dec 27 08:18:02 crc kubenswrapper[4934]: I1227 08:18:02.387587 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d" Dec 27 08:18:03 crc kubenswrapper[4934]: I1227 08:18:03.048347 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d"] Dec 27 08:18:03 crc kubenswrapper[4934]: I1227 08:18:03.992560 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d" event={"ID":"8af50e00-8101-4c06-b3bc-93fd48753773","Type":"ContainerStarted","Data":"1a646355df5c5f384ed2e3adbf3b07e0ddcc9824fc01d64f1727a0f78ce7a798"} Dec 27 08:18:05 crc kubenswrapper[4934]: I1227 08:18:05.009170 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d" event={"ID":"8af50e00-8101-4c06-b3bc-93fd48753773","Type":"ContainerStarted","Data":"8c9e16d9814c297f9dc8014b7cf464c52629776311b1300ff40f3c7f46834336"} Dec 27 08:18:05 crc kubenswrapper[4934]: I1227 08:18:05.051891 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d" podStartSLOduration=2.361403002 podStartE2EDuration="3.051864056s" podCreationTimestamp="2025-12-27 08:18:02 +0000 UTC" firstStartedPulling="2025-12-27 08:18:03.059172672 +0000 UTC m=+2143.879613266" lastFinishedPulling="2025-12-27 08:18:03.749633716 +0000 UTC m=+2144.570074320" observedRunningTime="2025-12-27 08:18:05.027675572 +0000 UTC m=+2145.848116186" watchObservedRunningTime="2025-12-27 08:18:05.051864056 +0000 UTC m=+2145.872304670" Dec 27 08:18:05 crc kubenswrapper[4934]: I1227 08:18:05.252144 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hhp5b" Dec 27 08:18:05 crc kubenswrapper[4934]: I1227 08:18:05.253446 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hhp5b" Dec 27 08:18:05 crc kubenswrapper[4934]: I1227 08:18:05.343443 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hhp5b" Dec 27 08:18:06 crc kubenswrapper[4934]: I1227 08:18:06.073921 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hhp5b" Dec 27 08:18:06 crc kubenswrapper[4934]: I1227 08:18:06.129506 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hhp5b"] Dec 27 08:18:08 crc kubenswrapper[4934]: I1227 08:18:08.055474 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hhp5b" podUID="8a018a29-f75d-47b4-9af4-e170e9110551" containerName="registry-server" containerID="cri-o://75dad1f790c0bb0acb68803def294aa97efe47b8cb06d9deba69c5f39d54028e" gracePeriod=2 Dec 27 08:18:08 crc kubenswrapper[4934]: I1227 08:18:08.607915 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hhp5b" Dec 27 08:18:08 crc kubenswrapper[4934]: I1227 08:18:08.713566 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ff27\" (UniqueName: \"kubernetes.io/projected/8a018a29-f75d-47b4-9af4-e170e9110551-kube-api-access-5ff27\") pod \"8a018a29-f75d-47b4-9af4-e170e9110551\" (UID: \"8a018a29-f75d-47b4-9af4-e170e9110551\") " Dec 27 08:18:08 crc kubenswrapper[4934]: I1227 08:18:08.713656 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a018a29-f75d-47b4-9af4-e170e9110551-catalog-content\") pod \"8a018a29-f75d-47b4-9af4-e170e9110551\" (UID: \"8a018a29-f75d-47b4-9af4-e170e9110551\") " Dec 27 08:18:08 crc kubenswrapper[4934]: I1227 08:18:08.713715 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a018a29-f75d-47b4-9af4-e170e9110551-utilities\") pod \"8a018a29-f75d-47b4-9af4-e170e9110551\" (UID: \"8a018a29-f75d-47b4-9af4-e170e9110551\") " Dec 27 08:18:08 crc kubenswrapper[4934]: I1227 08:18:08.714978 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a018a29-f75d-47b4-9af4-e170e9110551-utilities" (OuterVolumeSpecName: "utilities") pod "8a018a29-f75d-47b4-9af4-e170e9110551" (UID: "8a018a29-f75d-47b4-9af4-e170e9110551"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:18:08 crc kubenswrapper[4934]: I1227 08:18:08.720142 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a018a29-f75d-47b4-9af4-e170e9110551-kube-api-access-5ff27" (OuterVolumeSpecName: "kube-api-access-5ff27") pod "8a018a29-f75d-47b4-9af4-e170e9110551" (UID: "8a018a29-f75d-47b4-9af4-e170e9110551"). InnerVolumeSpecName "kube-api-access-5ff27". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:18:08 crc kubenswrapper[4934]: I1227 08:18:08.761411 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a018a29-f75d-47b4-9af4-e170e9110551-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a018a29-f75d-47b4-9af4-e170e9110551" (UID: "8a018a29-f75d-47b4-9af4-e170e9110551"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:18:08 crc kubenswrapper[4934]: I1227 08:18:08.816260 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ff27\" (UniqueName: \"kubernetes.io/projected/8a018a29-f75d-47b4-9af4-e170e9110551-kube-api-access-5ff27\") on node \"crc\" DevicePath \"\"" Dec 27 08:18:08 crc kubenswrapper[4934]: I1227 08:18:08.816291 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a018a29-f75d-47b4-9af4-e170e9110551-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 08:18:08 crc kubenswrapper[4934]: I1227 08:18:08.816301 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a018a29-f75d-47b4-9af4-e170e9110551-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 08:18:09 crc kubenswrapper[4934]: I1227 08:18:09.070132 4934 generic.go:334] "Generic (PLEG): container finished" podID="8a018a29-f75d-47b4-9af4-e170e9110551" containerID="75dad1f790c0bb0acb68803def294aa97efe47b8cb06d9deba69c5f39d54028e" exitCode=0 Dec 27 08:18:09 crc kubenswrapper[4934]: I1227 08:18:09.070177 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hhp5b" event={"ID":"8a018a29-f75d-47b4-9af4-e170e9110551","Type":"ContainerDied","Data":"75dad1f790c0bb0acb68803def294aa97efe47b8cb06d9deba69c5f39d54028e"} Dec 27 08:18:09 crc kubenswrapper[4934]: I1227 08:18:09.070191 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hhp5b" Dec 27 08:18:09 crc kubenswrapper[4934]: I1227 08:18:09.070216 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hhp5b" event={"ID":"8a018a29-f75d-47b4-9af4-e170e9110551","Type":"ContainerDied","Data":"6e6a1513b42af1b724c7a951def4877b11c251157bdcc688ff96ab305b5ece9d"} Dec 27 08:18:09 crc kubenswrapper[4934]: I1227 08:18:09.070238 4934 scope.go:117] "RemoveContainer" containerID="75dad1f790c0bb0acb68803def294aa97efe47b8cb06d9deba69c5f39d54028e" Dec 27 08:18:09 crc kubenswrapper[4934]: I1227 08:18:09.112498 4934 scope.go:117] "RemoveContainer" containerID="ab7022b7090c09020243fecf51234bb3388f36232d6f18cca805a0fda9fc515e" Dec 27 08:18:09 crc kubenswrapper[4934]: I1227 08:18:09.122233 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hhp5b"] Dec 27 08:18:09 crc kubenswrapper[4934]: I1227 08:18:09.134164 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hhp5b"] Dec 27 08:18:09 crc kubenswrapper[4934]: I1227 08:18:09.143687 4934 scope.go:117] "RemoveContainer" containerID="aec809f3a76f8f1291d6c7d892ad88fd4f7d8ab9242f76449ead528917ad28f0" Dec 27 08:18:09 crc kubenswrapper[4934]: I1227 08:18:09.188805 4934 scope.go:117] "RemoveContainer" containerID="75dad1f790c0bb0acb68803def294aa97efe47b8cb06d9deba69c5f39d54028e" Dec 27 08:18:09 crc kubenswrapper[4934]: E1227 08:18:09.189400 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75dad1f790c0bb0acb68803def294aa97efe47b8cb06d9deba69c5f39d54028e\": container with ID starting with 75dad1f790c0bb0acb68803def294aa97efe47b8cb06d9deba69c5f39d54028e not found: ID does not exist" containerID="75dad1f790c0bb0acb68803def294aa97efe47b8cb06d9deba69c5f39d54028e" Dec 27 08:18:09 crc kubenswrapper[4934]: I1227 08:18:09.189457 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75dad1f790c0bb0acb68803def294aa97efe47b8cb06d9deba69c5f39d54028e"} err="failed to get container status \"75dad1f790c0bb0acb68803def294aa97efe47b8cb06d9deba69c5f39d54028e\": rpc error: code = NotFound desc = could not find container \"75dad1f790c0bb0acb68803def294aa97efe47b8cb06d9deba69c5f39d54028e\": container with ID starting with 75dad1f790c0bb0acb68803def294aa97efe47b8cb06d9deba69c5f39d54028e not found: ID does not exist" Dec 27 08:18:09 crc kubenswrapper[4934]: I1227 08:18:09.189493 4934 scope.go:117] "RemoveContainer" containerID="ab7022b7090c09020243fecf51234bb3388f36232d6f18cca805a0fda9fc515e" Dec 27 08:18:09 crc kubenswrapper[4934]: E1227 08:18:09.190055 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab7022b7090c09020243fecf51234bb3388f36232d6f18cca805a0fda9fc515e\": container with ID starting with ab7022b7090c09020243fecf51234bb3388f36232d6f18cca805a0fda9fc515e not found: ID does not exist" containerID="ab7022b7090c09020243fecf51234bb3388f36232d6f18cca805a0fda9fc515e" Dec 27 08:18:09 crc kubenswrapper[4934]: I1227 08:18:09.190128 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab7022b7090c09020243fecf51234bb3388f36232d6f18cca805a0fda9fc515e"} err="failed to get container status \"ab7022b7090c09020243fecf51234bb3388f36232d6f18cca805a0fda9fc515e\": rpc error: code = NotFound desc = could not find container \"ab7022b7090c09020243fecf51234bb3388f36232d6f18cca805a0fda9fc515e\": container with ID starting with ab7022b7090c09020243fecf51234bb3388f36232d6f18cca805a0fda9fc515e not found: ID does not exist" Dec 27 08:18:09 crc kubenswrapper[4934]: I1227 08:18:09.190159 4934 scope.go:117] "RemoveContainer" containerID="aec809f3a76f8f1291d6c7d892ad88fd4f7d8ab9242f76449ead528917ad28f0" Dec 27 08:18:09 crc kubenswrapper[4934]: E1227 08:18:09.190686 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aec809f3a76f8f1291d6c7d892ad88fd4f7d8ab9242f76449ead528917ad28f0\": container with ID starting with aec809f3a76f8f1291d6c7d892ad88fd4f7d8ab9242f76449ead528917ad28f0 not found: ID does not exist" containerID="aec809f3a76f8f1291d6c7d892ad88fd4f7d8ab9242f76449ead528917ad28f0" Dec 27 08:18:09 crc kubenswrapper[4934]: I1227 08:18:09.190720 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aec809f3a76f8f1291d6c7d892ad88fd4f7d8ab9242f76449ead528917ad28f0"} err="failed to get container status \"aec809f3a76f8f1291d6c7d892ad88fd4f7d8ab9242f76449ead528917ad28f0\": rpc error: code = NotFound desc = could not find container \"aec809f3a76f8f1291d6c7d892ad88fd4f7d8ab9242f76449ead528917ad28f0\": container with ID starting with aec809f3a76f8f1291d6c7d892ad88fd4f7d8ab9242f76449ead528917ad28f0 not found: ID does not exist" Dec 27 08:18:09 crc kubenswrapper[4934]: I1227 08:18:09.484432 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a018a29-f75d-47b4-9af4-e170e9110551" path="/var/lib/kubelet/pods/8a018a29-f75d-47b4-9af4-e170e9110551/volumes" Dec 27 08:18:28 crc kubenswrapper[4934]: I1227 08:18:28.008092 4934 scope.go:117] "RemoveContainer" containerID="9a51d5582a010610a3618c11caa15ff0a3118518e903e82cfdf985039b49ef73" Dec 27 08:18:45 crc kubenswrapper[4934]: I1227 08:18:45.330269 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:18:45 crc kubenswrapper[4934]: I1227 08:18:45.330852 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:19:03 crc kubenswrapper[4934]: I1227 08:19:03.885556 4934 generic.go:334] "Generic (PLEG): container finished" podID="8af50e00-8101-4c06-b3bc-93fd48753773" containerID="8c9e16d9814c297f9dc8014b7cf464c52629776311b1300ff40f3c7f46834336" exitCode=0 Dec 27 08:19:03 crc kubenswrapper[4934]: I1227 08:19:03.885634 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d" event={"ID":"8af50e00-8101-4c06-b3bc-93fd48753773","Type":"ContainerDied","Data":"8c9e16d9814c297f9dc8014b7cf464c52629776311b1300ff40f3c7f46834336"} Dec 27 08:19:05 crc kubenswrapper[4934]: I1227 08:19:05.540201 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d" Dec 27 08:19:05 crc kubenswrapper[4934]: I1227 08:19:05.729221 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd8ft\" (UniqueName: \"kubernetes.io/projected/8af50e00-8101-4c06-b3bc-93fd48753773-kube-api-access-jd8ft\") pod \"8af50e00-8101-4c06-b3bc-93fd48753773\" (UID: \"8af50e00-8101-4c06-b3bc-93fd48753773\") " Dec 27 08:19:05 crc kubenswrapper[4934]: I1227 08:19:05.729817 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8af50e00-8101-4c06-b3bc-93fd48753773-ssh-key\") pod \"8af50e00-8101-4c06-b3bc-93fd48753773\" (UID: \"8af50e00-8101-4c06-b3bc-93fd48753773\") " Dec 27 08:19:05 crc kubenswrapper[4934]: I1227 08:19:05.729928 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8af50e00-8101-4c06-b3bc-93fd48753773-inventory\") pod \"8af50e00-8101-4c06-b3bc-93fd48753773\" (UID: \"8af50e00-8101-4c06-b3bc-93fd48753773\") " Dec 27 08:19:05 crc kubenswrapper[4934]: I1227 08:19:05.734038 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8af50e00-8101-4c06-b3bc-93fd48753773-kube-api-access-jd8ft" (OuterVolumeSpecName: "kube-api-access-jd8ft") pod "8af50e00-8101-4c06-b3bc-93fd48753773" (UID: "8af50e00-8101-4c06-b3bc-93fd48753773"). InnerVolumeSpecName "kube-api-access-jd8ft". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:19:05 crc kubenswrapper[4934]: I1227 08:19:05.759627 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8af50e00-8101-4c06-b3bc-93fd48753773-inventory" (OuterVolumeSpecName: "inventory") pod "8af50e00-8101-4c06-b3bc-93fd48753773" (UID: "8af50e00-8101-4c06-b3bc-93fd48753773"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:19:05 crc kubenswrapper[4934]: I1227 08:19:05.798217 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8af50e00-8101-4c06-b3bc-93fd48753773-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8af50e00-8101-4c06-b3bc-93fd48753773" (UID: "8af50e00-8101-4c06-b3bc-93fd48753773"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:19:05 crc kubenswrapper[4934]: I1227 08:19:05.833337 4934 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8af50e00-8101-4c06-b3bc-93fd48753773-inventory\") on node \"crc\" DevicePath \"\"" Dec 27 08:19:05 crc kubenswrapper[4934]: I1227 08:19:05.833404 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd8ft\" (UniqueName: \"kubernetes.io/projected/8af50e00-8101-4c06-b3bc-93fd48753773-kube-api-access-jd8ft\") on node \"crc\" DevicePath \"\"" Dec 27 08:19:05 crc kubenswrapper[4934]: I1227 08:19:05.833424 4934 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8af50e00-8101-4c06-b3bc-93fd48753773-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 27 08:19:05 crc kubenswrapper[4934]: I1227 08:19:05.913901 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d" event={"ID":"8af50e00-8101-4c06-b3bc-93fd48753773","Type":"ContainerDied","Data":"1a646355df5c5f384ed2e3adbf3b07e0ddcc9824fc01d64f1727a0f78ce7a798"} Dec 27 08:19:05 crc kubenswrapper[4934]: I1227 08:19:05.913944 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a646355df5c5f384ed2e3adbf3b07e0ddcc9824fc01d64f1727a0f78ce7a798" Dec 27 08:19:05 crc kubenswrapper[4934]: I1227 08:19:05.914006 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.030848 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-d6vnn"] Dec 27 08:19:06 crc kubenswrapper[4934]: E1227 08:19:06.031424 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a018a29-f75d-47b4-9af4-e170e9110551" containerName="registry-server" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.031450 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a018a29-f75d-47b4-9af4-e170e9110551" containerName="registry-server" Dec 27 08:19:06 crc kubenswrapper[4934]: E1227 08:19:06.031479 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a018a29-f75d-47b4-9af4-e170e9110551" containerName="extract-content" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.031490 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a018a29-f75d-47b4-9af4-e170e9110551" containerName="extract-content" Dec 27 08:19:06 crc kubenswrapper[4934]: E1227 08:19:06.031508 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8af50e00-8101-4c06-b3bc-93fd48753773" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.031517 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8af50e00-8101-4c06-b3bc-93fd48753773" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 27 08:19:06 crc kubenswrapper[4934]: E1227 08:19:06.031538 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a018a29-f75d-47b4-9af4-e170e9110551" containerName="extract-utilities" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.031546 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a018a29-f75d-47b4-9af4-e170e9110551" containerName="extract-utilities" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.031820 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8af50e00-8101-4c06-b3bc-93fd48753773" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.031854 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a018a29-f75d-47b4-9af4-e170e9110551" containerName="registry-server" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.032951 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-d6vnn" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.035316 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.035533 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.037336 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.039327 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-98ddn" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.050446 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-d6vnn"] Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.139174 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-d6vnn\" (UID: \"66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35\") " pod="openstack/ssh-known-hosts-edpm-deployment-d6vnn" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.139252 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-d6vnn\" (UID: \"66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35\") " pod="openstack/ssh-known-hosts-edpm-deployment-d6vnn" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.139357 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmnqh\" (UniqueName: \"kubernetes.io/projected/66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35-kube-api-access-fmnqh\") pod \"ssh-known-hosts-edpm-deployment-d6vnn\" (UID: \"66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35\") " pod="openstack/ssh-known-hosts-edpm-deployment-d6vnn" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.241495 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-d6vnn\" (UID: \"66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35\") " pod="openstack/ssh-known-hosts-edpm-deployment-d6vnn" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.241576 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-d6vnn\" (UID: \"66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35\") " pod="openstack/ssh-known-hosts-edpm-deployment-d6vnn" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.241704 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmnqh\" (UniqueName: \"kubernetes.io/projected/66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35-kube-api-access-fmnqh\") pod \"ssh-known-hosts-edpm-deployment-d6vnn\" (UID: \"66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35\") " pod="openstack/ssh-known-hosts-edpm-deployment-d6vnn" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.248200 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-d6vnn\" (UID: \"66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35\") " pod="openstack/ssh-known-hosts-edpm-deployment-d6vnn" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.248241 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-d6vnn\" (UID: \"66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35\") " pod="openstack/ssh-known-hosts-edpm-deployment-d6vnn" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.257609 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmnqh\" (UniqueName: \"kubernetes.io/projected/66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35-kube-api-access-fmnqh\") pod \"ssh-known-hosts-edpm-deployment-d6vnn\" (UID: \"66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35\") " pod="openstack/ssh-known-hosts-edpm-deployment-d6vnn" Dec 27 08:19:06 crc kubenswrapper[4934]: I1227 08:19:06.353172 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-d6vnn" Dec 27 08:19:07 crc kubenswrapper[4934]: I1227 08:19:07.121385 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-d6vnn"] Dec 27 08:19:07 crc kubenswrapper[4934]: I1227 08:19:07.945310 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-d6vnn" event={"ID":"66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35","Type":"ContainerStarted","Data":"b5a2c292a91989ca3755e4db66f0c0d1ad797b8557332638171475a03da0af03"} Dec 27 08:19:07 crc kubenswrapper[4934]: I1227 08:19:07.945879 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-d6vnn" event={"ID":"66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35","Type":"ContainerStarted","Data":"87c4aad828d5a5833aba4ce7b3845d7b92822238bafc8180ebddc88dcc1a3b7e"} Dec 27 08:19:07 crc kubenswrapper[4934]: I1227 08:19:07.978251 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-d6vnn" podStartSLOduration=1.5008286370000001 podStartE2EDuration="1.978235849s" podCreationTimestamp="2025-12-27 08:19:06 +0000 UTC" firstStartedPulling="2025-12-27 08:19:07.110245633 +0000 UTC m=+2207.930686227" lastFinishedPulling="2025-12-27 08:19:07.587652815 +0000 UTC m=+2208.408093439" observedRunningTime="2025-12-27 08:19:07.97465238 +0000 UTC m=+2208.795092974" watchObservedRunningTime="2025-12-27 08:19:07.978235849 +0000 UTC m=+2208.798676443" Dec 27 08:19:15 crc kubenswrapper[4934]: I1227 08:19:15.330337 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:19:15 crc kubenswrapper[4934]: I1227 08:19:15.331068 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:19:16 crc kubenswrapper[4934]: I1227 08:19:16.075345 4934 generic.go:334] "Generic (PLEG): container finished" podID="66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35" containerID="b5a2c292a91989ca3755e4db66f0c0d1ad797b8557332638171475a03da0af03" exitCode=0 Dec 27 08:19:16 crc kubenswrapper[4934]: I1227 08:19:16.075428 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-d6vnn" event={"ID":"66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35","Type":"ContainerDied","Data":"b5a2c292a91989ca3755e4db66f0c0d1ad797b8557332638171475a03da0af03"} Dec 27 08:19:17 crc kubenswrapper[4934]: I1227 08:19:17.587582 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-d6vnn" Dec 27 08:19:17 crc kubenswrapper[4934]: I1227 08:19:17.705545 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35-ssh-key-openstack-edpm-ipam\") pod \"66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35\" (UID: \"66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35\") " Dec 27 08:19:17 crc kubenswrapper[4934]: I1227 08:19:17.705937 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35-inventory-0\") pod \"66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35\" (UID: \"66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35\") " Dec 27 08:19:17 crc kubenswrapper[4934]: I1227 08:19:17.705983 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmnqh\" (UniqueName: \"kubernetes.io/projected/66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35-kube-api-access-fmnqh\") pod \"66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35\" (UID: \"66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35\") " Dec 27 08:19:17 crc kubenswrapper[4934]: I1227 08:19:17.710847 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35-kube-api-access-fmnqh" (OuterVolumeSpecName: "kube-api-access-fmnqh") pod "66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35" (UID: "66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35"). InnerVolumeSpecName "kube-api-access-fmnqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:19:17 crc kubenswrapper[4934]: I1227 08:19:17.741195 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35" (UID: "66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:19:17 crc kubenswrapper[4934]: I1227 08:19:17.771462 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35" (UID: "66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:19:17 crc kubenswrapper[4934]: I1227 08:19:17.808979 4934 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:19:17 crc kubenswrapper[4934]: I1227 08:19:17.809001 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmnqh\" (UniqueName: \"kubernetes.io/projected/66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35-kube-api-access-fmnqh\") on node \"crc\" DevicePath \"\"" Dec 27 08:19:17 crc kubenswrapper[4934]: I1227 08:19:17.809012 4934 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.103996 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-d6vnn" event={"ID":"66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35","Type":"ContainerDied","Data":"87c4aad828d5a5833aba4ce7b3845d7b92822238bafc8180ebddc88dcc1a3b7e"} Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.104056 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87c4aad828d5a5833aba4ce7b3845d7b92822238bafc8180ebddc88dcc1a3b7e" Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.104105 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-d6vnn" Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.210782 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9cms"] Dec 27 08:19:18 crc kubenswrapper[4934]: E1227 08:19:18.211664 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35" containerName="ssh-known-hosts-edpm-deployment" Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.211752 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35" containerName="ssh-known-hosts-edpm-deployment" Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.212217 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35" containerName="ssh-known-hosts-edpm-deployment" Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.213756 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9cms" Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.218072 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.218516 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.218599 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.218547 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-98ddn" Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.223569 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9cms"] Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.322434 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19b8530e-a7de-43d2-bfa7-b0f41ff55156-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j9cms\" (UID: \"19b8530e-a7de-43d2-bfa7-b0f41ff55156\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9cms" Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.322585 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19b8530e-a7de-43d2-bfa7-b0f41ff55156-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j9cms\" (UID: \"19b8530e-a7de-43d2-bfa7-b0f41ff55156\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9cms" Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.322834 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d9nm\" (UniqueName: \"kubernetes.io/projected/19b8530e-a7de-43d2-bfa7-b0f41ff55156-kube-api-access-8d9nm\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j9cms\" (UID: \"19b8530e-a7de-43d2-bfa7-b0f41ff55156\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9cms" Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.425293 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19b8530e-a7de-43d2-bfa7-b0f41ff55156-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j9cms\" (UID: \"19b8530e-a7de-43d2-bfa7-b0f41ff55156\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9cms" Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.425445 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19b8530e-a7de-43d2-bfa7-b0f41ff55156-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j9cms\" (UID: \"19b8530e-a7de-43d2-bfa7-b0f41ff55156\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9cms" Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.425526 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d9nm\" (UniqueName: \"kubernetes.io/projected/19b8530e-a7de-43d2-bfa7-b0f41ff55156-kube-api-access-8d9nm\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j9cms\" (UID: \"19b8530e-a7de-43d2-bfa7-b0f41ff55156\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9cms" Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.429176 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19b8530e-a7de-43d2-bfa7-b0f41ff55156-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j9cms\" (UID: \"19b8530e-a7de-43d2-bfa7-b0f41ff55156\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9cms" Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.429456 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19b8530e-a7de-43d2-bfa7-b0f41ff55156-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j9cms\" (UID: \"19b8530e-a7de-43d2-bfa7-b0f41ff55156\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9cms" Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.448348 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d9nm\" (UniqueName: \"kubernetes.io/projected/19b8530e-a7de-43d2-bfa7-b0f41ff55156-kube-api-access-8d9nm\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j9cms\" (UID: \"19b8530e-a7de-43d2-bfa7-b0f41ff55156\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9cms" Dec 27 08:19:18 crc kubenswrapper[4934]: I1227 08:19:18.558999 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9cms" Dec 27 08:19:19 crc kubenswrapper[4934]: I1227 08:19:19.209192 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9cms"] Dec 27 08:19:19 crc kubenswrapper[4934]: I1227 08:19:19.628144 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 27 08:19:20 crc kubenswrapper[4934]: I1227 08:19:20.057651 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-njzm7"] Dec 27 08:19:20 crc kubenswrapper[4934]: I1227 08:19:20.076722 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-njzm7"] Dec 27 08:19:20 crc kubenswrapper[4934]: I1227 08:19:20.132351 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9cms" event={"ID":"19b8530e-a7de-43d2-bfa7-b0f41ff55156","Type":"ContainerStarted","Data":"26b7e2e4ae627127d906d86c937e25675d6a06ebe68f5fef9e97a7ead3aa3959"} Dec 27 08:19:20 crc kubenswrapper[4934]: I1227 08:19:20.132405 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9cms" event={"ID":"19b8530e-a7de-43d2-bfa7-b0f41ff55156","Type":"ContainerStarted","Data":"c08147ded74b16c9845125c37701afb01b86610d1b1a5d7425650e510bf9e535"} Dec 27 08:19:20 crc kubenswrapper[4934]: I1227 08:19:20.149142 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9cms" podStartSLOduration=1.733075865 podStartE2EDuration="2.149125834s" podCreationTimestamp="2025-12-27 08:19:18 +0000 UTC" firstStartedPulling="2025-12-27 08:19:19.209610682 +0000 UTC m=+2220.030051276" lastFinishedPulling="2025-12-27 08:19:19.625660651 +0000 UTC m=+2220.446101245" observedRunningTime="2025-12-27 08:19:20.146758475 +0000 UTC m=+2220.967199069" watchObservedRunningTime="2025-12-27 08:19:20.149125834 +0000 UTC m=+2220.969566428" Dec 27 08:19:21 crc kubenswrapper[4934]: I1227 08:19:21.492506 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45927aae-ef6c-475a-b48f-59cd34ff1278" path="/var/lib/kubelet/pods/45927aae-ef6c-475a-b48f-59cd34ff1278/volumes" Dec 27 08:19:28 crc kubenswrapper[4934]: I1227 08:19:28.121159 4934 scope.go:117] "RemoveContainer" containerID="3a8c5484b16cfc1e465ec7b486a2604f58342498ccc50b8493e5c4578f77417d" Dec 27 08:19:29 crc kubenswrapper[4934]: I1227 08:19:29.243488 4934 generic.go:334] "Generic (PLEG): container finished" podID="19b8530e-a7de-43d2-bfa7-b0f41ff55156" containerID="26b7e2e4ae627127d906d86c937e25675d6a06ebe68f5fef9e97a7ead3aa3959" exitCode=0 Dec 27 08:19:29 crc kubenswrapper[4934]: I1227 08:19:29.243624 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9cms" event={"ID":"19b8530e-a7de-43d2-bfa7-b0f41ff55156","Type":"ContainerDied","Data":"26b7e2e4ae627127d906d86c937e25675d6a06ebe68f5fef9e97a7ead3aa3959"} Dec 27 08:19:30 crc kubenswrapper[4934]: I1227 08:19:30.849254 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9cms" Dec 27 08:19:30 crc kubenswrapper[4934]: I1227 08:19:30.999338 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19b8530e-a7de-43d2-bfa7-b0f41ff55156-inventory\") pod \"19b8530e-a7de-43d2-bfa7-b0f41ff55156\" (UID: \"19b8530e-a7de-43d2-bfa7-b0f41ff55156\") " Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:30.999969 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8d9nm\" (UniqueName: \"kubernetes.io/projected/19b8530e-a7de-43d2-bfa7-b0f41ff55156-kube-api-access-8d9nm\") pod \"19b8530e-a7de-43d2-bfa7-b0f41ff55156\" (UID: \"19b8530e-a7de-43d2-bfa7-b0f41ff55156\") " Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.000253 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19b8530e-a7de-43d2-bfa7-b0f41ff55156-ssh-key\") pod \"19b8530e-a7de-43d2-bfa7-b0f41ff55156\" (UID: \"19b8530e-a7de-43d2-bfa7-b0f41ff55156\") " Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.008831 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19b8530e-a7de-43d2-bfa7-b0f41ff55156-kube-api-access-8d9nm" (OuterVolumeSpecName: "kube-api-access-8d9nm") pod "19b8530e-a7de-43d2-bfa7-b0f41ff55156" (UID: "19b8530e-a7de-43d2-bfa7-b0f41ff55156"). InnerVolumeSpecName "kube-api-access-8d9nm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.037265 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19b8530e-a7de-43d2-bfa7-b0f41ff55156-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "19b8530e-a7de-43d2-bfa7-b0f41ff55156" (UID: "19b8530e-a7de-43d2-bfa7-b0f41ff55156"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.052442 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19b8530e-a7de-43d2-bfa7-b0f41ff55156-inventory" (OuterVolumeSpecName: "inventory") pod "19b8530e-a7de-43d2-bfa7-b0f41ff55156" (UID: "19b8530e-a7de-43d2-bfa7-b0f41ff55156"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.103172 4934 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19b8530e-a7de-43d2-bfa7-b0f41ff55156-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.103207 4934 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19b8530e-a7de-43d2-bfa7-b0f41ff55156-inventory\") on node \"crc\" DevicePath \"\"" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.103218 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8d9nm\" (UniqueName: \"kubernetes.io/projected/19b8530e-a7de-43d2-bfa7-b0f41ff55156-kube-api-access-8d9nm\") on node \"crc\" DevicePath \"\"" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.271526 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9cms" event={"ID":"19b8530e-a7de-43d2-bfa7-b0f41ff55156","Type":"ContainerDied","Data":"c08147ded74b16c9845125c37701afb01b86610d1b1a5d7425650e510bf9e535"} Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.271586 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c08147ded74b16c9845125c37701afb01b86610d1b1a5d7425650e510bf9e535" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.271648 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9cms" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.338665 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8"] Dec 27 08:19:31 crc kubenswrapper[4934]: E1227 08:19:31.339126 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19b8530e-a7de-43d2-bfa7-b0f41ff55156" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.339142 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="19b8530e-a7de-43d2-bfa7-b0f41ff55156" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.339343 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="19b8530e-a7de-43d2-bfa7-b0f41ff55156" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.340246 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.342821 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.343015 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.343675 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-98ddn" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.347975 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.356432 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8"] Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.514401 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5bs7\" (UniqueName: \"kubernetes.io/projected/a05cb14c-f210-4e43-8a1a-95fb4e86d55b-kube-api-access-d5bs7\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8\" (UID: \"a05cb14c-f210-4e43-8a1a-95fb4e86d55b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.514499 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a05cb14c-f210-4e43-8a1a-95fb4e86d55b-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8\" (UID: \"a05cb14c-f210-4e43-8a1a-95fb4e86d55b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.514689 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a05cb14c-f210-4e43-8a1a-95fb4e86d55b-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8\" (UID: \"a05cb14c-f210-4e43-8a1a-95fb4e86d55b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.617053 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a05cb14c-f210-4e43-8a1a-95fb4e86d55b-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8\" (UID: \"a05cb14c-f210-4e43-8a1a-95fb4e86d55b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.617257 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5bs7\" (UniqueName: \"kubernetes.io/projected/a05cb14c-f210-4e43-8a1a-95fb4e86d55b-kube-api-access-d5bs7\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8\" (UID: \"a05cb14c-f210-4e43-8a1a-95fb4e86d55b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.617357 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a05cb14c-f210-4e43-8a1a-95fb4e86d55b-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8\" (UID: \"a05cb14c-f210-4e43-8a1a-95fb4e86d55b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.624036 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a05cb14c-f210-4e43-8a1a-95fb4e86d55b-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8\" (UID: \"a05cb14c-f210-4e43-8a1a-95fb4e86d55b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.629335 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a05cb14c-f210-4e43-8a1a-95fb4e86d55b-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8\" (UID: \"a05cb14c-f210-4e43-8a1a-95fb4e86d55b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.655144 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5bs7\" (UniqueName: \"kubernetes.io/projected/a05cb14c-f210-4e43-8a1a-95fb4e86d55b-kube-api-access-d5bs7\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8\" (UID: \"a05cb14c-f210-4e43-8a1a-95fb4e86d55b\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8" Dec 27 08:19:31 crc kubenswrapper[4934]: I1227 08:19:31.665281 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8" Dec 27 08:19:32 crc kubenswrapper[4934]: I1227 08:19:32.286012 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8"] Dec 27 08:19:32 crc kubenswrapper[4934]: I1227 08:19:32.766954 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-74zvl"] Dec 27 08:19:32 crc kubenswrapper[4934]: I1227 08:19:32.770277 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-74zvl" Dec 27 08:19:32 crc kubenswrapper[4934]: I1227 08:19:32.785663 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-74zvl"] Dec 27 08:19:32 crc kubenswrapper[4934]: I1227 08:19:32.958681 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90ec261f-2792-4a52-85fd-3d7073f8a6ce-utilities\") pod \"redhat-marketplace-74zvl\" (UID: \"90ec261f-2792-4a52-85fd-3d7073f8a6ce\") " pod="openshift-marketplace/redhat-marketplace-74zvl" Dec 27 08:19:32 crc kubenswrapper[4934]: I1227 08:19:32.959035 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsczt\" (UniqueName: \"kubernetes.io/projected/90ec261f-2792-4a52-85fd-3d7073f8a6ce-kube-api-access-xsczt\") pod \"redhat-marketplace-74zvl\" (UID: \"90ec261f-2792-4a52-85fd-3d7073f8a6ce\") " pod="openshift-marketplace/redhat-marketplace-74zvl" Dec 27 08:19:32 crc kubenswrapper[4934]: I1227 08:19:32.959377 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90ec261f-2792-4a52-85fd-3d7073f8a6ce-catalog-content\") pod \"redhat-marketplace-74zvl\" (UID: \"90ec261f-2792-4a52-85fd-3d7073f8a6ce\") " pod="openshift-marketplace/redhat-marketplace-74zvl" Dec 27 08:19:33 crc kubenswrapper[4934]: I1227 08:19:33.061689 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90ec261f-2792-4a52-85fd-3d7073f8a6ce-catalog-content\") pod \"redhat-marketplace-74zvl\" (UID: \"90ec261f-2792-4a52-85fd-3d7073f8a6ce\") " pod="openshift-marketplace/redhat-marketplace-74zvl" Dec 27 08:19:33 crc kubenswrapper[4934]: I1227 08:19:33.061901 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90ec261f-2792-4a52-85fd-3d7073f8a6ce-utilities\") pod \"redhat-marketplace-74zvl\" (UID: \"90ec261f-2792-4a52-85fd-3d7073f8a6ce\") " pod="openshift-marketplace/redhat-marketplace-74zvl" Dec 27 08:19:33 crc kubenswrapper[4934]: I1227 08:19:33.062032 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsczt\" (UniqueName: \"kubernetes.io/projected/90ec261f-2792-4a52-85fd-3d7073f8a6ce-kube-api-access-xsczt\") pod \"redhat-marketplace-74zvl\" (UID: \"90ec261f-2792-4a52-85fd-3d7073f8a6ce\") " pod="openshift-marketplace/redhat-marketplace-74zvl" Dec 27 08:19:33 crc kubenswrapper[4934]: I1227 08:19:33.062780 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90ec261f-2792-4a52-85fd-3d7073f8a6ce-catalog-content\") pod \"redhat-marketplace-74zvl\" (UID: \"90ec261f-2792-4a52-85fd-3d7073f8a6ce\") " pod="openshift-marketplace/redhat-marketplace-74zvl" Dec 27 08:19:33 crc kubenswrapper[4934]: I1227 08:19:33.062807 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90ec261f-2792-4a52-85fd-3d7073f8a6ce-utilities\") pod \"redhat-marketplace-74zvl\" (UID: \"90ec261f-2792-4a52-85fd-3d7073f8a6ce\") " pod="openshift-marketplace/redhat-marketplace-74zvl" Dec 27 08:19:33 crc kubenswrapper[4934]: I1227 08:19:33.079351 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsczt\" (UniqueName: \"kubernetes.io/projected/90ec261f-2792-4a52-85fd-3d7073f8a6ce-kube-api-access-xsczt\") pod \"redhat-marketplace-74zvl\" (UID: \"90ec261f-2792-4a52-85fd-3d7073f8a6ce\") " pod="openshift-marketplace/redhat-marketplace-74zvl" Dec 27 08:19:33 crc kubenswrapper[4934]: I1227 08:19:33.102519 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-74zvl" Dec 27 08:19:33 crc kubenswrapper[4934]: I1227 08:19:33.300609 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8" event={"ID":"a05cb14c-f210-4e43-8a1a-95fb4e86d55b","Type":"ContainerStarted","Data":"fff50e09a0632fa272f642b60ec89ac7d0c25437c33d2aa1160ab821a6db7c39"} Dec 27 08:19:33 crc kubenswrapper[4934]: I1227 08:19:33.300664 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8" event={"ID":"a05cb14c-f210-4e43-8a1a-95fb4e86d55b","Type":"ContainerStarted","Data":"4453d835877321ea2d2daa8fcf5135aab14efef1b6461a5eafcaa357168ec393"} Dec 27 08:19:33 crc kubenswrapper[4934]: I1227 08:19:33.318467 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8" podStartSLOduration=1.7734740119999999 podStartE2EDuration="2.318448492s" podCreationTimestamp="2025-12-27 08:19:31 +0000 UTC" firstStartedPulling="2025-12-27 08:19:32.290258885 +0000 UTC m=+2233.110699489" lastFinishedPulling="2025-12-27 08:19:32.835233375 +0000 UTC m=+2233.655673969" observedRunningTime="2025-12-27 08:19:33.315753675 +0000 UTC m=+2234.136194299" watchObservedRunningTime="2025-12-27 08:19:33.318448492 +0000 UTC m=+2234.138889086" Dec 27 08:19:33 crc kubenswrapper[4934]: I1227 08:19:33.606593 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-74zvl"] Dec 27 08:19:34 crc kubenswrapper[4934]: I1227 08:19:34.317337 4934 generic.go:334] "Generic (PLEG): container finished" podID="90ec261f-2792-4a52-85fd-3d7073f8a6ce" containerID="37176a8f8cf15d1b097e36cc5cc04cded57bb6c31346cdb98e9303e252d2fd3b" exitCode=0 Dec 27 08:19:34 crc kubenswrapper[4934]: I1227 08:19:34.317456 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-74zvl" event={"ID":"90ec261f-2792-4a52-85fd-3d7073f8a6ce","Type":"ContainerDied","Data":"37176a8f8cf15d1b097e36cc5cc04cded57bb6c31346cdb98e9303e252d2fd3b"} Dec 27 08:19:34 crc kubenswrapper[4934]: I1227 08:19:34.317847 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-74zvl" event={"ID":"90ec261f-2792-4a52-85fd-3d7073f8a6ce","Type":"ContainerStarted","Data":"b36188876c397fd16922d66769a3242c118c64d9945dd868ede16d0472bd72c6"} Dec 27 08:19:35 crc kubenswrapper[4934]: I1227 08:19:35.336681 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-74zvl" event={"ID":"90ec261f-2792-4a52-85fd-3d7073f8a6ce","Type":"ContainerStarted","Data":"4207a1a7a656f6a3175b3e9df481ec19aed4a88e58d76cedc95ca5ff514c1fd0"} Dec 27 08:19:36 crc kubenswrapper[4934]: I1227 08:19:36.355044 4934 generic.go:334] "Generic (PLEG): container finished" podID="90ec261f-2792-4a52-85fd-3d7073f8a6ce" containerID="4207a1a7a656f6a3175b3e9df481ec19aed4a88e58d76cedc95ca5ff514c1fd0" exitCode=0 Dec 27 08:19:36 crc kubenswrapper[4934]: I1227 08:19:36.355330 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-74zvl" event={"ID":"90ec261f-2792-4a52-85fd-3d7073f8a6ce","Type":"ContainerDied","Data":"4207a1a7a656f6a3175b3e9df481ec19aed4a88e58d76cedc95ca5ff514c1fd0"} Dec 27 08:19:37 crc kubenswrapper[4934]: I1227 08:19:37.371772 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-74zvl" event={"ID":"90ec261f-2792-4a52-85fd-3d7073f8a6ce","Type":"ContainerStarted","Data":"c991c00ee4dfe899ea5134b6398e6108cd8da3f23c3139a1a19683d01e04fcff"} Dec 27 08:19:37 crc kubenswrapper[4934]: I1227 08:19:37.405510 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-74zvl" podStartSLOduration=2.839249241 podStartE2EDuration="5.405492938s" podCreationTimestamp="2025-12-27 08:19:32 +0000 UTC" firstStartedPulling="2025-12-27 08:19:34.32024227 +0000 UTC m=+2235.140682894" lastFinishedPulling="2025-12-27 08:19:36.886485997 +0000 UTC m=+2237.706926591" observedRunningTime="2025-12-27 08:19:37.396940564 +0000 UTC m=+2238.217381158" watchObservedRunningTime="2025-12-27 08:19:37.405492938 +0000 UTC m=+2238.225933532" Dec 27 08:19:39 crc kubenswrapper[4934]: I1227 08:19:39.338812 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-klw9h"] Dec 27 08:19:39 crc kubenswrapper[4934]: I1227 08:19:39.342265 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-klw9h" Dec 27 08:19:39 crc kubenswrapper[4934]: I1227 08:19:39.354098 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-klw9h"] Dec 27 08:19:39 crc kubenswrapper[4934]: I1227 08:19:39.437862 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2723cec-4aa9-488f-ad6b-2da5239e9d58-utilities\") pod \"redhat-operators-klw9h\" (UID: \"a2723cec-4aa9-488f-ad6b-2da5239e9d58\") " pod="openshift-marketplace/redhat-operators-klw9h" Dec 27 08:19:39 crc kubenswrapper[4934]: I1227 08:19:39.437969 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2723cec-4aa9-488f-ad6b-2da5239e9d58-catalog-content\") pod \"redhat-operators-klw9h\" (UID: \"a2723cec-4aa9-488f-ad6b-2da5239e9d58\") " pod="openshift-marketplace/redhat-operators-klw9h" Dec 27 08:19:39 crc kubenswrapper[4934]: I1227 08:19:39.438188 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spbv5\" (UniqueName: \"kubernetes.io/projected/a2723cec-4aa9-488f-ad6b-2da5239e9d58-kube-api-access-spbv5\") pod \"redhat-operators-klw9h\" (UID: \"a2723cec-4aa9-488f-ad6b-2da5239e9d58\") " pod="openshift-marketplace/redhat-operators-klw9h" Dec 27 08:19:39 crc kubenswrapper[4934]: I1227 08:19:39.541094 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2723cec-4aa9-488f-ad6b-2da5239e9d58-utilities\") pod \"redhat-operators-klw9h\" (UID: \"a2723cec-4aa9-488f-ad6b-2da5239e9d58\") " pod="openshift-marketplace/redhat-operators-klw9h" Dec 27 08:19:39 crc kubenswrapper[4934]: I1227 08:19:39.541410 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2723cec-4aa9-488f-ad6b-2da5239e9d58-catalog-content\") pod \"redhat-operators-klw9h\" (UID: \"a2723cec-4aa9-488f-ad6b-2da5239e9d58\") " pod="openshift-marketplace/redhat-operators-klw9h" Dec 27 08:19:39 crc kubenswrapper[4934]: I1227 08:19:39.541540 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spbv5\" (UniqueName: \"kubernetes.io/projected/a2723cec-4aa9-488f-ad6b-2da5239e9d58-kube-api-access-spbv5\") pod \"redhat-operators-klw9h\" (UID: \"a2723cec-4aa9-488f-ad6b-2da5239e9d58\") " pod="openshift-marketplace/redhat-operators-klw9h" Dec 27 08:19:39 crc kubenswrapper[4934]: I1227 08:19:39.541642 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2723cec-4aa9-488f-ad6b-2da5239e9d58-utilities\") pod \"redhat-operators-klw9h\" (UID: \"a2723cec-4aa9-488f-ad6b-2da5239e9d58\") " pod="openshift-marketplace/redhat-operators-klw9h" Dec 27 08:19:39 crc kubenswrapper[4934]: I1227 08:19:39.541884 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2723cec-4aa9-488f-ad6b-2da5239e9d58-catalog-content\") pod \"redhat-operators-klw9h\" (UID: \"a2723cec-4aa9-488f-ad6b-2da5239e9d58\") " pod="openshift-marketplace/redhat-operators-klw9h" Dec 27 08:19:39 crc kubenswrapper[4934]: I1227 08:19:39.564864 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spbv5\" (UniqueName: \"kubernetes.io/projected/a2723cec-4aa9-488f-ad6b-2da5239e9d58-kube-api-access-spbv5\") pod \"redhat-operators-klw9h\" (UID: \"a2723cec-4aa9-488f-ad6b-2da5239e9d58\") " pod="openshift-marketplace/redhat-operators-klw9h" Dec 27 08:19:39 crc kubenswrapper[4934]: I1227 08:19:39.665481 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-klw9h" Dec 27 08:19:40 crc kubenswrapper[4934]: I1227 08:19:40.221650 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-klw9h"] Dec 27 08:19:40 crc kubenswrapper[4934]: W1227 08:19:40.228918 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2723cec_4aa9_488f_ad6b_2da5239e9d58.slice/crio-6500c7ee8c7308d353b2aef33052d5a725a217ddbc8718f797b002f906c1a087 WatchSource:0}: Error finding container 6500c7ee8c7308d353b2aef33052d5a725a217ddbc8718f797b002f906c1a087: Status 404 returned error can't find the container with id 6500c7ee8c7308d353b2aef33052d5a725a217ddbc8718f797b002f906c1a087 Dec 27 08:19:40 crc kubenswrapper[4934]: I1227 08:19:40.406452 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-klw9h" event={"ID":"a2723cec-4aa9-488f-ad6b-2da5239e9d58","Type":"ContainerStarted","Data":"6500c7ee8c7308d353b2aef33052d5a725a217ddbc8718f797b002f906c1a087"} Dec 27 08:19:41 crc kubenswrapper[4934]: I1227 08:19:41.419632 4934 generic.go:334] "Generic (PLEG): container finished" podID="a2723cec-4aa9-488f-ad6b-2da5239e9d58" containerID="e0705d07aab475678dc5532014f201df2549bd289bc082132354c4ea5a4c4e6a" exitCode=0 Dec 27 08:19:41 crc kubenswrapper[4934]: I1227 08:19:41.419812 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-klw9h" event={"ID":"a2723cec-4aa9-488f-ad6b-2da5239e9d58","Type":"ContainerDied","Data":"e0705d07aab475678dc5532014f201df2549bd289bc082132354c4ea5a4c4e6a"} Dec 27 08:19:42 crc kubenswrapper[4934]: I1227 08:19:42.435675 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-klw9h" event={"ID":"a2723cec-4aa9-488f-ad6b-2da5239e9d58","Type":"ContainerStarted","Data":"2e05d297cc807570d51219fede4e77c028fb9386d554d6b02c7b30117fb307ca"} Dec 27 08:19:43 crc kubenswrapper[4934]: I1227 08:19:43.104068 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-74zvl" Dec 27 08:19:43 crc kubenswrapper[4934]: I1227 08:19:43.104488 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-74zvl" Dec 27 08:19:43 crc kubenswrapper[4934]: I1227 08:19:43.162645 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-74zvl" Dec 27 08:19:43 crc kubenswrapper[4934]: I1227 08:19:43.449952 4934 generic.go:334] "Generic (PLEG): container finished" podID="a05cb14c-f210-4e43-8a1a-95fb4e86d55b" containerID="fff50e09a0632fa272f642b60ec89ac7d0c25437c33d2aa1160ab821a6db7c39" exitCode=0 Dec 27 08:19:43 crc kubenswrapper[4934]: I1227 08:19:43.450055 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8" event={"ID":"a05cb14c-f210-4e43-8a1a-95fb4e86d55b","Type":"ContainerDied","Data":"fff50e09a0632fa272f642b60ec89ac7d0c25437c33d2aa1160ab821a6db7c39"} Dec 27 08:19:43 crc kubenswrapper[4934]: I1227 08:19:43.523906 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-74zvl" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.087613 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.284818 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a05cb14c-f210-4e43-8a1a-95fb4e86d55b-ssh-key\") pod \"a05cb14c-f210-4e43-8a1a-95fb4e86d55b\" (UID: \"a05cb14c-f210-4e43-8a1a-95fb4e86d55b\") " Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.285283 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5bs7\" (UniqueName: \"kubernetes.io/projected/a05cb14c-f210-4e43-8a1a-95fb4e86d55b-kube-api-access-d5bs7\") pod \"a05cb14c-f210-4e43-8a1a-95fb4e86d55b\" (UID: \"a05cb14c-f210-4e43-8a1a-95fb4e86d55b\") " Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.285450 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a05cb14c-f210-4e43-8a1a-95fb4e86d55b-inventory\") pod \"a05cb14c-f210-4e43-8a1a-95fb4e86d55b\" (UID: \"a05cb14c-f210-4e43-8a1a-95fb4e86d55b\") " Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.290230 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a05cb14c-f210-4e43-8a1a-95fb4e86d55b-kube-api-access-d5bs7" (OuterVolumeSpecName: "kube-api-access-d5bs7") pod "a05cb14c-f210-4e43-8a1a-95fb4e86d55b" (UID: "a05cb14c-f210-4e43-8a1a-95fb4e86d55b"). InnerVolumeSpecName "kube-api-access-d5bs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.314363 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-74zvl"] Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.328357 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a05cb14c-f210-4e43-8a1a-95fb4e86d55b-inventory" (OuterVolumeSpecName: "inventory") pod "a05cb14c-f210-4e43-8a1a-95fb4e86d55b" (UID: "a05cb14c-f210-4e43-8a1a-95fb4e86d55b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.330631 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.330682 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.330727 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.331655 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.331709 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" gracePeriod=600 Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.350818 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a05cb14c-f210-4e43-8a1a-95fb4e86d55b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a05cb14c-f210-4e43-8a1a-95fb4e86d55b" (UID: "a05cb14c-f210-4e43-8a1a-95fb4e86d55b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.388612 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5bs7\" (UniqueName: \"kubernetes.io/projected/a05cb14c-f210-4e43-8a1a-95fb4e86d55b-kube-api-access-d5bs7\") on node \"crc\" DevicePath \"\"" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.388645 4934 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a05cb14c-f210-4e43-8a1a-95fb4e86d55b-inventory\") on node \"crc\" DevicePath \"\"" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.388654 4934 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a05cb14c-f210-4e43-8a1a-95fb4e86d55b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.479149 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.483837 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" exitCode=0 Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.484059 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-74zvl" podUID="90ec261f-2792-4a52-85fd-3d7073f8a6ce" containerName="registry-server" containerID="cri-o://c991c00ee4dfe899ea5134b6398e6108cd8da3f23c3139a1a19683d01e04fcff" gracePeriod=2 Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.485012 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8" event={"ID":"a05cb14c-f210-4e43-8a1a-95fb4e86d55b","Type":"ContainerDied","Data":"4453d835877321ea2d2daa8fcf5135aab14efef1b6461a5eafcaa357168ec393"} Dec 27 08:19:45 crc kubenswrapper[4934]: E1227 08:19:45.485130 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.485173 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4453d835877321ea2d2daa8fcf5135aab14efef1b6461a5eafcaa357168ec393" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.485214 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6"} Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.485265 4934 scope.go:117] "RemoveContainer" containerID="deb6fb2d2d95a934db9a65eebdf5e58dc64e13629a44e74765a4e64756d050bb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.617754 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb"] Dec 27 08:19:45 crc kubenswrapper[4934]: E1227 08:19:45.618290 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a05cb14c-f210-4e43-8a1a-95fb4e86d55b" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.618308 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a05cb14c-f210-4e43-8a1a-95fb4e86d55b" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.618521 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a05cb14c-f210-4e43-8a1a-95fb4e86d55b" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.619380 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.622327 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.622374 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.622549 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.622697 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.622872 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.623005 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.623074 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.623258 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-98ddn" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.623485 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.632541 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb"] Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.805291 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.805811 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.805923 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.806322 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.806387 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.806465 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.806502 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.806551 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.806740 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.806811 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.807072 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.807147 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.807191 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.807235 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.807323 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpvvq\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-kube-api-access-wpvvq\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.807356 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.909403 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.909474 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.909519 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.909543 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.909573 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.909602 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.909638 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpvvq\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-kube-api-access-wpvvq\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.909661 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.909701 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.909801 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.909842 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.909911 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.909940 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.909975 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.910000 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.910027 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.915264 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.915548 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.917149 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.917240 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.917734 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.917808 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.918891 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.919286 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.920584 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.921622 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.921809 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.921902 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.923629 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.924747 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.937432 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.940793 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpvvq\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-kube-api-access-wpvvq\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:45 crc kubenswrapper[4934]: I1227 08:19:45.948002 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.071828 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-74zvl" Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.118204 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90ec261f-2792-4a52-85fd-3d7073f8a6ce-utilities\") pod \"90ec261f-2792-4a52-85fd-3d7073f8a6ce\" (UID: \"90ec261f-2792-4a52-85fd-3d7073f8a6ce\") " Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.118279 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsczt\" (UniqueName: \"kubernetes.io/projected/90ec261f-2792-4a52-85fd-3d7073f8a6ce-kube-api-access-xsczt\") pod \"90ec261f-2792-4a52-85fd-3d7073f8a6ce\" (UID: \"90ec261f-2792-4a52-85fd-3d7073f8a6ce\") " Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.118305 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90ec261f-2792-4a52-85fd-3d7073f8a6ce-catalog-content\") pod \"90ec261f-2792-4a52-85fd-3d7073f8a6ce\" (UID: \"90ec261f-2792-4a52-85fd-3d7073f8a6ce\") " Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.121235 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90ec261f-2792-4a52-85fd-3d7073f8a6ce-utilities" (OuterVolumeSpecName: "utilities") pod "90ec261f-2792-4a52-85fd-3d7073f8a6ce" (UID: "90ec261f-2792-4a52-85fd-3d7073f8a6ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.124507 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90ec261f-2792-4a52-85fd-3d7073f8a6ce-kube-api-access-xsczt" (OuterVolumeSpecName: "kube-api-access-xsczt") pod "90ec261f-2792-4a52-85fd-3d7073f8a6ce" (UID: "90ec261f-2792-4a52-85fd-3d7073f8a6ce"). InnerVolumeSpecName "kube-api-access-xsczt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.229681 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90ec261f-2792-4a52-85fd-3d7073f8a6ce-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.229716 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsczt\" (UniqueName: \"kubernetes.io/projected/90ec261f-2792-4a52-85fd-3d7073f8a6ce-kube-api-access-xsczt\") on node \"crc\" DevicePath \"\"" Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.233651 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90ec261f-2792-4a52-85fd-3d7073f8a6ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "90ec261f-2792-4a52-85fd-3d7073f8a6ce" (UID: "90ec261f-2792-4a52-85fd-3d7073f8a6ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.333807 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90ec261f-2792-4a52-85fd-3d7073f8a6ce-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.496602 4934 generic.go:334] "Generic (PLEG): container finished" podID="90ec261f-2792-4a52-85fd-3d7073f8a6ce" containerID="c991c00ee4dfe899ea5134b6398e6108cd8da3f23c3139a1a19683d01e04fcff" exitCode=0 Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.496669 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-74zvl" event={"ID":"90ec261f-2792-4a52-85fd-3d7073f8a6ce","Type":"ContainerDied","Data":"c991c00ee4dfe899ea5134b6398e6108cd8da3f23c3139a1a19683d01e04fcff"} Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.496686 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-74zvl" Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.496709 4934 scope.go:117] "RemoveContainer" containerID="c991c00ee4dfe899ea5134b6398e6108cd8da3f23c3139a1a19683d01e04fcff" Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.496696 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-74zvl" event={"ID":"90ec261f-2792-4a52-85fd-3d7073f8a6ce","Type":"ContainerDied","Data":"b36188876c397fd16922d66769a3242c118c64d9945dd868ede16d0472bd72c6"} Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.507428 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:19:46 crc kubenswrapper[4934]: E1227 08:19:46.507739 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.510578 4934 generic.go:334] "Generic (PLEG): container finished" podID="a2723cec-4aa9-488f-ad6b-2da5239e9d58" containerID="2e05d297cc807570d51219fede4e77c028fb9386d554d6b02c7b30117fb307ca" exitCode=0 Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.510625 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-klw9h" event={"ID":"a2723cec-4aa9-488f-ad6b-2da5239e9d58","Type":"ContainerDied","Data":"2e05d297cc807570d51219fede4e77c028fb9386d554d6b02c7b30117fb307ca"} Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.538362 4934 scope.go:117] "RemoveContainer" containerID="4207a1a7a656f6a3175b3e9df481ec19aed4a88e58d76cedc95ca5ff514c1fd0" Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.565265 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-74zvl"] Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.580862 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-74zvl"] Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.593908 4934 scope.go:117] "RemoveContainer" containerID="37176a8f8cf15d1b097e36cc5cc04cded57bb6c31346cdb98e9303e252d2fd3b" Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.632129 4934 scope.go:117] "RemoveContainer" containerID="c991c00ee4dfe899ea5134b6398e6108cd8da3f23c3139a1a19683d01e04fcff" Dec 27 08:19:46 crc kubenswrapper[4934]: E1227 08:19:46.632660 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c991c00ee4dfe899ea5134b6398e6108cd8da3f23c3139a1a19683d01e04fcff\": container with ID starting with c991c00ee4dfe899ea5134b6398e6108cd8da3f23c3139a1a19683d01e04fcff not found: ID does not exist" containerID="c991c00ee4dfe899ea5134b6398e6108cd8da3f23c3139a1a19683d01e04fcff" Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.632691 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c991c00ee4dfe899ea5134b6398e6108cd8da3f23c3139a1a19683d01e04fcff"} err="failed to get container status \"c991c00ee4dfe899ea5134b6398e6108cd8da3f23c3139a1a19683d01e04fcff\": rpc error: code = NotFound desc = could not find container \"c991c00ee4dfe899ea5134b6398e6108cd8da3f23c3139a1a19683d01e04fcff\": container with ID starting with c991c00ee4dfe899ea5134b6398e6108cd8da3f23c3139a1a19683d01e04fcff not found: ID does not exist" Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.632712 4934 scope.go:117] "RemoveContainer" containerID="4207a1a7a656f6a3175b3e9df481ec19aed4a88e58d76cedc95ca5ff514c1fd0" Dec 27 08:19:46 crc kubenswrapper[4934]: E1227 08:19:46.632967 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4207a1a7a656f6a3175b3e9df481ec19aed4a88e58d76cedc95ca5ff514c1fd0\": container with ID starting with 4207a1a7a656f6a3175b3e9df481ec19aed4a88e58d76cedc95ca5ff514c1fd0 not found: ID does not exist" containerID="4207a1a7a656f6a3175b3e9df481ec19aed4a88e58d76cedc95ca5ff514c1fd0" Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.633005 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4207a1a7a656f6a3175b3e9df481ec19aed4a88e58d76cedc95ca5ff514c1fd0"} err="failed to get container status \"4207a1a7a656f6a3175b3e9df481ec19aed4a88e58d76cedc95ca5ff514c1fd0\": rpc error: code = NotFound desc = could not find container \"4207a1a7a656f6a3175b3e9df481ec19aed4a88e58d76cedc95ca5ff514c1fd0\": container with ID starting with 4207a1a7a656f6a3175b3e9df481ec19aed4a88e58d76cedc95ca5ff514c1fd0 not found: ID does not exist" Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.633030 4934 scope.go:117] "RemoveContainer" containerID="37176a8f8cf15d1b097e36cc5cc04cded57bb6c31346cdb98e9303e252d2fd3b" Dec 27 08:19:46 crc kubenswrapper[4934]: E1227 08:19:46.633482 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37176a8f8cf15d1b097e36cc5cc04cded57bb6c31346cdb98e9303e252d2fd3b\": container with ID starting with 37176a8f8cf15d1b097e36cc5cc04cded57bb6c31346cdb98e9303e252d2fd3b not found: ID does not exist" containerID="37176a8f8cf15d1b097e36cc5cc04cded57bb6c31346cdb98e9303e252d2fd3b" Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.633505 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37176a8f8cf15d1b097e36cc5cc04cded57bb6c31346cdb98e9303e252d2fd3b"} err="failed to get container status \"37176a8f8cf15d1b097e36cc5cc04cded57bb6c31346cdb98e9303e252d2fd3b\": rpc error: code = NotFound desc = could not find container \"37176a8f8cf15d1b097e36cc5cc04cded57bb6c31346cdb98e9303e252d2fd3b\": container with ID starting with 37176a8f8cf15d1b097e36cc5cc04cded57bb6c31346cdb98e9303e252d2fd3b not found: ID does not exist" Dec 27 08:19:46 crc kubenswrapper[4934]: I1227 08:19:46.659977 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb"] Dec 27 08:19:47 crc kubenswrapper[4934]: I1227 08:19:47.484482 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90ec261f-2792-4a52-85fd-3d7073f8a6ce" path="/var/lib/kubelet/pods/90ec261f-2792-4a52-85fd-3d7073f8a6ce/volumes" Dec 27 08:19:47 crc kubenswrapper[4934]: I1227 08:19:47.522627 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-klw9h" event={"ID":"a2723cec-4aa9-488f-ad6b-2da5239e9d58","Type":"ContainerStarted","Data":"29d089cc2a3fcd142f8282977cb2d70d74746c8252132d4a37053dae945f5676"} Dec 27 08:19:47 crc kubenswrapper[4934]: I1227 08:19:47.526638 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" event={"ID":"a5cad0ac-88e1-47e4-8539-da835534e9c5","Type":"ContainerStarted","Data":"f96126779bfe481620bcb4d3339f456249d66dcaebdc02ebabe6220227a4e55b"} Dec 27 08:19:47 crc kubenswrapper[4934]: I1227 08:19:47.526679 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" event={"ID":"a5cad0ac-88e1-47e4-8539-da835534e9c5","Type":"ContainerStarted","Data":"8dfd3f888f43ce22d8e62ef288137cebbcfef0dc4023df304c9065392fb1c8cb"} Dec 27 08:19:47 crc kubenswrapper[4934]: I1227 08:19:47.527269 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:19:47 crc kubenswrapper[4934]: E1227 08:19:47.527538 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:19:47 crc kubenswrapper[4934]: I1227 08:19:47.552692 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-klw9h" podStartSLOduration=2.930750357 podStartE2EDuration="8.552667904s" podCreationTimestamp="2025-12-27 08:19:39 +0000 UTC" firstStartedPulling="2025-12-27 08:19:41.421284984 +0000 UTC m=+2242.241725578" lastFinishedPulling="2025-12-27 08:19:47.043202501 +0000 UTC m=+2247.863643125" observedRunningTime="2025-12-27 08:19:47.537104456 +0000 UTC m=+2248.357545070" watchObservedRunningTime="2025-12-27 08:19:47.552667904 +0000 UTC m=+2248.373108508" Dec 27 08:19:47 crc kubenswrapper[4934]: I1227 08:19:47.565779 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" podStartSLOduration=2.142600224 podStartE2EDuration="2.565757691s" podCreationTimestamp="2025-12-27 08:19:45 +0000 UTC" firstStartedPulling="2025-12-27 08:19:46.688513644 +0000 UTC m=+2247.508954238" lastFinishedPulling="2025-12-27 08:19:47.111671101 +0000 UTC m=+2247.932111705" observedRunningTime="2025-12-27 08:19:47.560166662 +0000 UTC m=+2248.380607276" watchObservedRunningTime="2025-12-27 08:19:47.565757691 +0000 UTC m=+2248.386198285" Dec 27 08:19:49 crc kubenswrapper[4934]: I1227 08:19:49.666132 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-klw9h" Dec 27 08:19:49 crc kubenswrapper[4934]: I1227 08:19:49.667417 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-klw9h" Dec 27 08:19:50 crc kubenswrapper[4934]: I1227 08:19:50.718816 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-klw9h" podUID="a2723cec-4aa9-488f-ad6b-2da5239e9d58" containerName="registry-server" probeResult="failure" output=< Dec 27 08:19:50 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 08:19:50 crc kubenswrapper[4934]: > Dec 27 08:20:00 crc kubenswrapper[4934]: I1227 08:20:00.061471 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-wgpjv"] Dec 27 08:20:00 crc kubenswrapper[4934]: I1227 08:20:00.077339 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-wgpjv"] Dec 27 08:20:00 crc kubenswrapper[4934]: I1227 08:20:00.722508 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-klw9h" podUID="a2723cec-4aa9-488f-ad6b-2da5239e9d58" containerName="registry-server" probeResult="failure" output=< Dec 27 08:20:00 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 08:20:00 crc kubenswrapper[4934]: > Dec 27 08:20:01 crc kubenswrapper[4934]: I1227 08:20:01.491567 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f05689e-1351-435f-a267-f0c6508f1f4a" path="/var/lib/kubelet/pods/4f05689e-1351-435f-a267-f0c6508f1f4a/volumes" Dec 27 08:20:02 crc kubenswrapper[4934]: I1227 08:20:02.468661 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:20:02 crc kubenswrapper[4934]: E1227 08:20:02.469594 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:20:09 crc kubenswrapper[4934]: I1227 08:20:09.727194 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-klw9h" Dec 27 08:20:09 crc kubenswrapper[4934]: I1227 08:20:09.785539 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-klw9h" Dec 27 08:20:11 crc kubenswrapper[4934]: I1227 08:20:11.778708 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-klw9h"] Dec 27 08:20:11 crc kubenswrapper[4934]: I1227 08:20:11.780857 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-klw9h" podUID="a2723cec-4aa9-488f-ad6b-2da5239e9d58" containerName="registry-server" containerID="cri-o://29d089cc2a3fcd142f8282977cb2d70d74746c8252132d4a37053dae945f5676" gracePeriod=2 Dec 27 08:20:12 crc kubenswrapper[4934]: I1227 08:20:12.302918 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-klw9h" Dec 27 08:20:12 crc kubenswrapper[4934]: I1227 08:20:12.392940 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spbv5\" (UniqueName: \"kubernetes.io/projected/a2723cec-4aa9-488f-ad6b-2da5239e9d58-kube-api-access-spbv5\") pod \"a2723cec-4aa9-488f-ad6b-2da5239e9d58\" (UID: \"a2723cec-4aa9-488f-ad6b-2da5239e9d58\") " Dec 27 08:20:12 crc kubenswrapper[4934]: I1227 08:20:12.393168 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2723cec-4aa9-488f-ad6b-2da5239e9d58-utilities\") pod \"a2723cec-4aa9-488f-ad6b-2da5239e9d58\" (UID: \"a2723cec-4aa9-488f-ad6b-2da5239e9d58\") " Dec 27 08:20:12 crc kubenswrapper[4934]: I1227 08:20:12.393247 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2723cec-4aa9-488f-ad6b-2da5239e9d58-catalog-content\") pod \"a2723cec-4aa9-488f-ad6b-2da5239e9d58\" (UID: \"a2723cec-4aa9-488f-ad6b-2da5239e9d58\") " Dec 27 08:20:12 crc kubenswrapper[4934]: I1227 08:20:12.394278 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2723cec-4aa9-488f-ad6b-2da5239e9d58-utilities" (OuterVolumeSpecName: "utilities") pod "a2723cec-4aa9-488f-ad6b-2da5239e9d58" (UID: "a2723cec-4aa9-488f-ad6b-2da5239e9d58"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:20:12 crc kubenswrapper[4934]: I1227 08:20:12.398491 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2723cec-4aa9-488f-ad6b-2da5239e9d58-kube-api-access-spbv5" (OuterVolumeSpecName: "kube-api-access-spbv5") pod "a2723cec-4aa9-488f-ad6b-2da5239e9d58" (UID: "a2723cec-4aa9-488f-ad6b-2da5239e9d58"). InnerVolumeSpecName "kube-api-access-spbv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:20:12 crc kubenswrapper[4934]: I1227 08:20:12.496423 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2723cec-4aa9-488f-ad6b-2da5239e9d58-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 08:20:12 crc kubenswrapper[4934]: I1227 08:20:12.496460 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spbv5\" (UniqueName: \"kubernetes.io/projected/a2723cec-4aa9-488f-ad6b-2da5239e9d58-kube-api-access-spbv5\") on node \"crc\" DevicePath \"\"" Dec 27 08:20:12 crc kubenswrapper[4934]: I1227 08:20:12.505966 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2723cec-4aa9-488f-ad6b-2da5239e9d58-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a2723cec-4aa9-488f-ad6b-2da5239e9d58" (UID: "a2723cec-4aa9-488f-ad6b-2da5239e9d58"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:20:12 crc kubenswrapper[4934]: I1227 08:20:12.599214 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2723cec-4aa9-488f-ad6b-2da5239e9d58-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 08:20:12 crc kubenswrapper[4934]: I1227 08:20:12.865070 4934 generic.go:334] "Generic (PLEG): container finished" podID="a2723cec-4aa9-488f-ad6b-2da5239e9d58" containerID="29d089cc2a3fcd142f8282977cb2d70d74746c8252132d4a37053dae945f5676" exitCode=0 Dec 27 08:20:12 crc kubenswrapper[4934]: I1227 08:20:12.865131 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-klw9h" event={"ID":"a2723cec-4aa9-488f-ad6b-2da5239e9d58","Type":"ContainerDied","Data":"29d089cc2a3fcd142f8282977cb2d70d74746c8252132d4a37053dae945f5676"} Dec 27 08:20:12 crc kubenswrapper[4934]: I1227 08:20:12.865164 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-klw9h" event={"ID":"a2723cec-4aa9-488f-ad6b-2da5239e9d58","Type":"ContainerDied","Data":"6500c7ee8c7308d353b2aef33052d5a725a217ddbc8718f797b002f906c1a087"} Dec 27 08:20:12 crc kubenswrapper[4934]: I1227 08:20:12.865175 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-klw9h" Dec 27 08:20:12 crc kubenswrapper[4934]: I1227 08:20:12.865204 4934 scope.go:117] "RemoveContainer" containerID="29d089cc2a3fcd142f8282977cb2d70d74746c8252132d4a37053dae945f5676" Dec 27 08:20:12 crc kubenswrapper[4934]: I1227 08:20:12.910545 4934 scope.go:117] "RemoveContainer" containerID="2e05d297cc807570d51219fede4e77c028fb9386d554d6b02c7b30117fb307ca" Dec 27 08:20:12 crc kubenswrapper[4934]: I1227 08:20:12.913070 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-klw9h"] Dec 27 08:20:12 crc kubenswrapper[4934]: I1227 08:20:12.947282 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-klw9h"] Dec 27 08:20:12 crc kubenswrapper[4934]: I1227 08:20:12.955561 4934 scope.go:117] "RemoveContainer" containerID="e0705d07aab475678dc5532014f201df2549bd289bc082132354c4ea5a4c4e6a" Dec 27 08:20:13 crc kubenswrapper[4934]: I1227 08:20:13.005827 4934 scope.go:117] "RemoveContainer" containerID="29d089cc2a3fcd142f8282977cb2d70d74746c8252132d4a37053dae945f5676" Dec 27 08:20:13 crc kubenswrapper[4934]: E1227 08:20:13.006672 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29d089cc2a3fcd142f8282977cb2d70d74746c8252132d4a37053dae945f5676\": container with ID starting with 29d089cc2a3fcd142f8282977cb2d70d74746c8252132d4a37053dae945f5676 not found: ID does not exist" containerID="29d089cc2a3fcd142f8282977cb2d70d74746c8252132d4a37053dae945f5676" Dec 27 08:20:13 crc kubenswrapper[4934]: I1227 08:20:13.006727 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29d089cc2a3fcd142f8282977cb2d70d74746c8252132d4a37053dae945f5676"} err="failed to get container status \"29d089cc2a3fcd142f8282977cb2d70d74746c8252132d4a37053dae945f5676\": rpc error: code = NotFound desc = could not find container \"29d089cc2a3fcd142f8282977cb2d70d74746c8252132d4a37053dae945f5676\": container with ID starting with 29d089cc2a3fcd142f8282977cb2d70d74746c8252132d4a37053dae945f5676 not found: ID does not exist" Dec 27 08:20:13 crc kubenswrapper[4934]: I1227 08:20:13.006765 4934 scope.go:117] "RemoveContainer" containerID="2e05d297cc807570d51219fede4e77c028fb9386d554d6b02c7b30117fb307ca" Dec 27 08:20:13 crc kubenswrapper[4934]: E1227 08:20:13.007310 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e05d297cc807570d51219fede4e77c028fb9386d554d6b02c7b30117fb307ca\": container with ID starting with 2e05d297cc807570d51219fede4e77c028fb9386d554d6b02c7b30117fb307ca not found: ID does not exist" containerID="2e05d297cc807570d51219fede4e77c028fb9386d554d6b02c7b30117fb307ca" Dec 27 08:20:13 crc kubenswrapper[4934]: I1227 08:20:13.013298 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e05d297cc807570d51219fede4e77c028fb9386d554d6b02c7b30117fb307ca"} err="failed to get container status \"2e05d297cc807570d51219fede4e77c028fb9386d554d6b02c7b30117fb307ca\": rpc error: code = NotFound desc = could not find container \"2e05d297cc807570d51219fede4e77c028fb9386d554d6b02c7b30117fb307ca\": container with ID starting with 2e05d297cc807570d51219fede4e77c028fb9386d554d6b02c7b30117fb307ca not found: ID does not exist" Dec 27 08:20:13 crc kubenswrapper[4934]: I1227 08:20:13.013339 4934 scope.go:117] "RemoveContainer" containerID="e0705d07aab475678dc5532014f201df2549bd289bc082132354c4ea5a4c4e6a" Dec 27 08:20:13 crc kubenswrapper[4934]: E1227 08:20:13.014696 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0705d07aab475678dc5532014f201df2549bd289bc082132354c4ea5a4c4e6a\": container with ID starting with e0705d07aab475678dc5532014f201df2549bd289bc082132354c4ea5a4c4e6a not found: ID does not exist" containerID="e0705d07aab475678dc5532014f201df2549bd289bc082132354c4ea5a4c4e6a" Dec 27 08:20:13 crc kubenswrapper[4934]: I1227 08:20:13.014770 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0705d07aab475678dc5532014f201df2549bd289bc082132354c4ea5a4c4e6a"} err="failed to get container status \"e0705d07aab475678dc5532014f201df2549bd289bc082132354c4ea5a4c4e6a\": rpc error: code = NotFound desc = could not find container \"e0705d07aab475678dc5532014f201df2549bd289bc082132354c4ea5a4c4e6a\": container with ID starting with e0705d07aab475678dc5532014f201df2549bd289bc082132354c4ea5a4c4e6a not found: ID does not exist" Dec 27 08:20:13 crc kubenswrapper[4934]: I1227 08:20:13.486304 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2723cec-4aa9-488f-ad6b-2da5239e9d58" path="/var/lib/kubelet/pods/a2723cec-4aa9-488f-ad6b-2da5239e9d58/volumes" Dec 27 08:20:14 crc kubenswrapper[4934]: I1227 08:20:14.468267 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:20:14 crc kubenswrapper[4934]: E1227 08:20:14.468796 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:20:15 crc kubenswrapper[4934]: I1227 08:20:15.616006 4934 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","poda05cb14c-f210-4e43-8a1a-95fb4e86d55b"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort poda05cb14c-f210-4e43-8a1a-95fb4e86d55b] : Timed out while waiting for systemd to remove kubepods-besteffort-poda05cb14c_f210_4e43_8a1a_95fb4e86d55b.slice" Dec 27 08:20:15 crc kubenswrapper[4934]: E1227 08:20:15.616074 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort poda05cb14c-f210-4e43-8a1a-95fb4e86d55b] : unable to destroy cgroup paths for cgroup [kubepods besteffort poda05cb14c-f210-4e43-8a1a-95fb4e86d55b] : Timed out while waiting for systemd to remove kubepods-besteffort-poda05cb14c_f210_4e43_8a1a_95fb4e86d55b.slice" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8" podUID="a05cb14c-f210-4e43-8a1a-95fb4e86d55b" Dec 27 08:20:15 crc kubenswrapper[4934]: I1227 08:20:15.932459 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8" Dec 27 08:20:27 crc kubenswrapper[4934]: I1227 08:20:27.468390 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:20:27 crc kubenswrapper[4934]: E1227 08:20:27.469653 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:20:28 crc kubenswrapper[4934]: I1227 08:20:28.217821 4934 scope.go:117] "RemoveContainer" containerID="d4029fef32d31d19e82de9403dc39d655329d6235f402b4de52d8250f613f1f0" Dec 27 08:20:39 crc kubenswrapper[4934]: I1227 08:20:39.266963 4934 generic.go:334] "Generic (PLEG): container finished" podID="a5cad0ac-88e1-47e4-8539-da835534e9c5" containerID="f96126779bfe481620bcb4d3339f456249d66dcaebdc02ebabe6220227a4e55b" exitCode=0 Dec 27 08:20:39 crc kubenswrapper[4934]: I1227 08:20:39.267656 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" event={"ID":"a5cad0ac-88e1-47e4-8539-da835534e9c5","Type":"ContainerDied","Data":"f96126779bfe481620bcb4d3339f456249d66dcaebdc02ebabe6220227a4e55b"} Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.467673 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:20:40 crc kubenswrapper[4934]: E1227 08:20:40.468388 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.807227 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.817599 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"a5cad0ac-88e1-47e4-8539-da835534e9c5\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.817723 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-telemetry-combined-ca-bundle\") pod \"a5cad0ac-88e1-47e4-8539-da835534e9c5\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.817753 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-ovn-default-certs-0\") pod \"a5cad0ac-88e1-47e4-8539-da835534e9c5\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.817789 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-telemetry-power-monitoring-combined-ca-bundle\") pod \"a5cad0ac-88e1-47e4-8539-da835534e9c5\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.817821 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpvvq\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-kube-api-access-wpvvq\") pod \"a5cad0ac-88e1-47e4-8539-da835534e9c5\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.817844 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-libvirt-combined-ca-bundle\") pod \"a5cad0ac-88e1-47e4-8539-da835534e9c5\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.817892 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-inventory\") pod \"a5cad0ac-88e1-47e4-8539-da835534e9c5\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.818542 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-bootstrap-combined-ca-bundle\") pod \"a5cad0ac-88e1-47e4-8539-da835534e9c5\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.818593 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-ovn-combined-ca-bundle\") pod \"a5cad0ac-88e1-47e4-8539-da835534e9c5\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.818628 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-neutron-metadata-combined-ca-bundle\") pod \"a5cad0ac-88e1-47e4-8539-da835534e9c5\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.818654 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"a5cad0ac-88e1-47e4-8539-da835534e9c5\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.818673 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-repo-setup-combined-ca-bundle\") pod \"a5cad0ac-88e1-47e4-8539-da835534e9c5\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.823561 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "a5cad0ac-88e1-47e4-8539-da835534e9c5" (UID: "a5cad0ac-88e1-47e4-8539-da835534e9c5"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.824618 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "a5cad0ac-88e1-47e4-8539-da835534e9c5" (UID: "a5cad0ac-88e1-47e4-8539-da835534e9c5"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.827895 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "a5cad0ac-88e1-47e4-8539-da835534e9c5" (UID: "a5cad0ac-88e1-47e4-8539-da835534e9c5"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.828437 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "a5cad0ac-88e1-47e4-8539-da835534e9c5" (UID: "a5cad0ac-88e1-47e4-8539-da835534e9c5"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.828489 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "a5cad0ac-88e1-47e4-8539-da835534e9c5" (UID: "a5cad0ac-88e1-47e4-8539-da835534e9c5"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.828604 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "a5cad0ac-88e1-47e4-8539-da835534e9c5" (UID: "a5cad0ac-88e1-47e4-8539-da835534e9c5"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.828652 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "a5cad0ac-88e1-47e4-8539-da835534e9c5" (UID: "a5cad0ac-88e1-47e4-8539-da835534e9c5"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.832675 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "a5cad0ac-88e1-47e4-8539-da835534e9c5" (UID: "a5cad0ac-88e1-47e4-8539-da835534e9c5"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.835319 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "a5cad0ac-88e1-47e4-8539-da835534e9c5" (UID: "a5cad0ac-88e1-47e4-8539-da835534e9c5"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.837070 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "a5cad0ac-88e1-47e4-8539-da835534e9c5" (UID: "a5cad0ac-88e1-47e4-8539-da835534e9c5"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.837317 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-kube-api-access-wpvvq" (OuterVolumeSpecName: "kube-api-access-wpvvq") pod "a5cad0ac-88e1-47e4-8539-da835534e9c5" (UID: "a5cad0ac-88e1-47e4-8539-da835534e9c5"). InnerVolumeSpecName "kube-api-access-wpvvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.856805 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-inventory" (OuterVolumeSpecName: "inventory") pod "a5cad0ac-88e1-47e4-8539-da835534e9c5" (UID: "a5cad0ac-88e1-47e4-8539-da835534e9c5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.920142 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"a5cad0ac-88e1-47e4-8539-da835534e9c5\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.920219 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-ssh-key\") pod \"a5cad0ac-88e1-47e4-8539-da835534e9c5\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.920290 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"a5cad0ac-88e1-47e4-8539-da835534e9c5\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.920317 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-nova-combined-ca-bundle\") pod \"a5cad0ac-88e1-47e4-8539-da835534e9c5\" (UID: \"a5cad0ac-88e1-47e4-8539-da835534e9c5\") " Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.921062 4934 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.921079 4934 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.921113 4934 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.921123 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpvvq\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-kube-api-access-wpvvq\") on node \"crc\" DevicePath \"\"" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.921133 4934 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.921142 4934 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-inventory\") on node \"crc\" DevicePath \"\"" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.921150 4934 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.921160 4934 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.921169 4934 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.921179 4934 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.921189 4934 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.921198 4934 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.925690 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "a5cad0ac-88e1-47e4-8539-da835534e9c5" (UID: "a5cad0ac-88e1-47e4-8539-da835534e9c5"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.925892 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "a5cad0ac-88e1-47e4-8539-da835534e9c5" (UID: "a5cad0ac-88e1-47e4-8539-da835534e9c5"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.926522 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0") pod "a5cad0ac-88e1-47e4-8539-da835534e9c5" (UID: "a5cad0ac-88e1-47e4-8539-da835534e9c5"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:20:40 crc kubenswrapper[4934]: I1227 08:20:40.949397 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a5cad0ac-88e1-47e4-8539-da835534e9c5" (UID: "a5cad0ac-88e1-47e4-8539-da835534e9c5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.023267 4934 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.023439 4934 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.023508 4934 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a5cad0ac-88e1-47e4-8539-da835534e9c5-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.023573 4934 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5cad0ac-88e1-47e4-8539-da835534e9c5-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.312188 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" event={"ID":"a5cad0ac-88e1-47e4-8539-da835534e9c5","Type":"ContainerDied","Data":"8dfd3f888f43ce22d8e62ef288137cebbcfef0dc4023df304c9065392fb1c8cb"} Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.312230 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8dfd3f888f43ce22d8e62ef288137cebbcfef0dc4023df304c9065392fb1c8cb" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.312247 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.393532 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp"] Dec 27 08:20:41 crc kubenswrapper[4934]: E1227 08:20:41.394153 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2723cec-4aa9-488f-ad6b-2da5239e9d58" containerName="registry-server" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.394178 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2723cec-4aa9-488f-ad6b-2da5239e9d58" containerName="registry-server" Dec 27 08:20:41 crc kubenswrapper[4934]: E1227 08:20:41.394208 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2723cec-4aa9-488f-ad6b-2da5239e9d58" containerName="extract-utilities" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.394217 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2723cec-4aa9-488f-ad6b-2da5239e9d58" containerName="extract-utilities" Dec 27 08:20:41 crc kubenswrapper[4934]: E1227 08:20:41.394240 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2723cec-4aa9-488f-ad6b-2da5239e9d58" containerName="extract-content" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.394248 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2723cec-4aa9-488f-ad6b-2da5239e9d58" containerName="extract-content" Dec 27 08:20:41 crc kubenswrapper[4934]: E1227 08:20:41.394263 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5cad0ac-88e1-47e4-8539-da835534e9c5" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.394271 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5cad0ac-88e1-47e4-8539-da835534e9c5" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 27 08:20:41 crc kubenswrapper[4934]: E1227 08:20:41.394291 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90ec261f-2792-4a52-85fd-3d7073f8a6ce" containerName="extract-content" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.394299 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="90ec261f-2792-4a52-85fd-3d7073f8a6ce" containerName="extract-content" Dec 27 08:20:41 crc kubenswrapper[4934]: E1227 08:20:41.394327 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90ec261f-2792-4a52-85fd-3d7073f8a6ce" containerName="registry-server" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.394335 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="90ec261f-2792-4a52-85fd-3d7073f8a6ce" containerName="registry-server" Dec 27 08:20:41 crc kubenswrapper[4934]: E1227 08:20:41.394353 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90ec261f-2792-4a52-85fd-3d7073f8a6ce" containerName="extract-utilities" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.394360 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="90ec261f-2792-4a52-85fd-3d7073f8a6ce" containerName="extract-utilities" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.394612 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="90ec261f-2792-4a52-85fd-3d7073f8a6ce" containerName="registry-server" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.394628 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2723cec-4aa9-488f-ad6b-2da5239e9d58" containerName="registry-server" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.394655 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5cad0ac-88e1-47e4-8539-da835534e9c5" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.396715 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.398902 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-98ddn" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.399115 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.399299 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.399591 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.399809 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.411048 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp"] Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.540591 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcbkd\" (UniqueName: \"kubernetes.io/projected/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-kube-api-access-zcbkd\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fcwdp\" (UID: \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.540804 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fcwdp\" (UID: \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.540938 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fcwdp\" (UID: \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.541040 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fcwdp\" (UID: \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.541163 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fcwdp\" (UID: \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.643953 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcbkd\" (UniqueName: \"kubernetes.io/projected/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-kube-api-access-zcbkd\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fcwdp\" (UID: \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.644046 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fcwdp\" (UID: \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.644074 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fcwdp\" (UID: \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.644117 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fcwdp\" (UID: \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.644150 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fcwdp\" (UID: \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.645021 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fcwdp\" (UID: \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.647238 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fcwdp\" (UID: \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.647712 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fcwdp\" (UID: \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.647723 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fcwdp\" (UID: \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.660182 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcbkd\" (UniqueName: \"kubernetes.io/projected/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-kube-api-access-zcbkd\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-fcwdp\" (UID: \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" Dec 27 08:20:41 crc kubenswrapper[4934]: I1227 08:20:41.724592 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" Dec 27 08:20:42 crc kubenswrapper[4934]: I1227 08:20:42.362342 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp"] Dec 27 08:20:43 crc kubenswrapper[4934]: I1227 08:20:43.339463 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" event={"ID":"595c2188-f2bb-4c19-b1a9-2da10bdc89c7","Type":"ContainerStarted","Data":"8f38ec4842a3f75a946e0437e962adf6950744abc9488af1e30a4b63ef82ccff"} Dec 27 08:20:43 crc kubenswrapper[4934]: I1227 08:20:43.341148 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" event={"ID":"595c2188-f2bb-4c19-b1a9-2da10bdc89c7","Type":"ContainerStarted","Data":"890b1eac3c0ab775dacf935c9e743054bce3a24c9878daba6b983458370a8d4c"} Dec 27 08:20:43 crc kubenswrapper[4934]: I1227 08:20:43.366662 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" podStartSLOduration=1.843664799 podStartE2EDuration="2.366636212s" podCreationTimestamp="2025-12-27 08:20:41 +0000 UTC" firstStartedPulling="2025-12-27 08:20:42.362887954 +0000 UTC m=+2303.183328548" lastFinishedPulling="2025-12-27 08:20:42.885859367 +0000 UTC m=+2303.706299961" observedRunningTime="2025-12-27 08:20:43.361020032 +0000 UTC m=+2304.181460626" watchObservedRunningTime="2025-12-27 08:20:43.366636212 +0000 UTC m=+2304.187076826" Dec 27 08:20:55 crc kubenswrapper[4934]: I1227 08:20:55.468327 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:20:55 crc kubenswrapper[4934]: E1227 08:20:55.469713 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:21:10 crc kubenswrapper[4934]: I1227 08:21:10.467137 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:21:10 crc kubenswrapper[4934]: E1227 08:21:10.467923 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:21:21 crc kubenswrapper[4934]: I1227 08:21:21.468729 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:21:21 crc kubenswrapper[4934]: E1227 08:21:21.478166 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:21:32 crc kubenswrapper[4934]: I1227 08:21:32.468669 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:21:32 crc kubenswrapper[4934]: E1227 08:21:32.469973 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:21:45 crc kubenswrapper[4934]: I1227 08:21:45.468806 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:21:45 crc kubenswrapper[4934]: E1227 08:21:45.469700 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:21:56 crc kubenswrapper[4934]: I1227 08:21:56.339860 4934 generic.go:334] "Generic (PLEG): container finished" podID="595c2188-f2bb-4c19-b1a9-2da10bdc89c7" containerID="8f38ec4842a3f75a946e0437e962adf6950744abc9488af1e30a4b63ef82ccff" exitCode=0 Dec 27 08:21:56 crc kubenswrapper[4934]: I1227 08:21:56.339930 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" event={"ID":"595c2188-f2bb-4c19-b1a9-2da10bdc89c7","Type":"ContainerDied","Data":"8f38ec4842a3f75a946e0437e962adf6950744abc9488af1e30a4b63ef82ccff"} Dec 27 08:21:57 crc kubenswrapper[4934]: I1227 08:21:57.472824 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:21:57 crc kubenswrapper[4934]: E1227 08:21:57.473678 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:21:57 crc kubenswrapper[4934]: I1227 08:21:57.910289 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.028153 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-ssh-key\") pod \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\" (UID: \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\") " Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.028207 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcbkd\" (UniqueName: \"kubernetes.io/projected/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-kube-api-access-zcbkd\") pod \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\" (UID: \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\") " Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.028229 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-inventory\") pod \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\" (UID: \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\") " Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.028308 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-ovn-combined-ca-bundle\") pod \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\" (UID: \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\") " Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.028483 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-ovncontroller-config-0\") pod \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\" (UID: \"595c2188-f2bb-4c19-b1a9-2da10bdc89c7\") " Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.033989 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "595c2188-f2bb-4c19-b1a9-2da10bdc89c7" (UID: "595c2188-f2bb-4c19-b1a9-2da10bdc89c7"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.034609 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-kube-api-access-zcbkd" (OuterVolumeSpecName: "kube-api-access-zcbkd") pod "595c2188-f2bb-4c19-b1a9-2da10bdc89c7" (UID: "595c2188-f2bb-4c19-b1a9-2da10bdc89c7"). InnerVolumeSpecName "kube-api-access-zcbkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.065701 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-inventory" (OuterVolumeSpecName: "inventory") pod "595c2188-f2bb-4c19-b1a9-2da10bdc89c7" (UID: "595c2188-f2bb-4c19-b1a9-2da10bdc89c7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.066235 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "595c2188-f2bb-4c19-b1a9-2da10bdc89c7" (UID: "595c2188-f2bb-4c19-b1a9-2da10bdc89c7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.078036 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "595c2188-f2bb-4c19-b1a9-2da10bdc89c7" (UID: "595c2188-f2bb-4c19-b1a9-2da10bdc89c7"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.131225 4934 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.131266 4934 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.131279 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcbkd\" (UniqueName: \"kubernetes.io/projected/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-kube-api-access-zcbkd\") on node \"crc\" DevicePath \"\"" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.131292 4934 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-inventory\") on node \"crc\" DevicePath \"\"" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.131304 4934 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/595c2188-f2bb-4c19-b1a9-2da10bdc89c7-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.367598 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" event={"ID":"595c2188-f2bb-4c19-b1a9-2da10bdc89c7","Type":"ContainerDied","Data":"890b1eac3c0ab775dacf935c9e743054bce3a24c9878daba6b983458370a8d4c"} Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.368025 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="890b1eac3c0ab775dacf935c9e743054bce3a24c9878daba6b983458370a8d4c" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.367689 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-fcwdp" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.502673 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk"] Dec 27 08:21:58 crc kubenswrapper[4934]: E1227 08:21:58.503585 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="595c2188-f2bb-4c19-b1a9-2da10bdc89c7" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.503613 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="595c2188-f2bb-4c19-b1a9-2da10bdc89c7" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.504164 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="595c2188-f2bb-4c19-b1a9-2da10bdc89c7" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.505755 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.512104 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.512255 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.512285 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.512449 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.512573 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-98ddn" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.512689 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.531847 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk"] Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.644582 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.644907 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2wh2\" (UniqueName: \"kubernetes.io/projected/34927e46-8eca-4b2d-ba3a-e68690b804df-kube-api-access-c2wh2\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.645120 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.645550 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.645866 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.645941 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.748681 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2wh2\" (UniqueName: \"kubernetes.io/projected/34927e46-8eca-4b2d-ba3a-e68690b804df-kube-api-access-c2wh2\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.748759 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.748872 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.748967 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.748995 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.749071 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.753777 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.754366 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.756673 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.756746 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.757215 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.772576 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2wh2\" (UniqueName: \"kubernetes.io/projected/34927e46-8eca-4b2d-ba3a-e68690b804df-kube-api-access-c2wh2\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:21:58 crc kubenswrapper[4934]: I1227 08:21:58.839616 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:21:59 crc kubenswrapper[4934]: I1227 08:21:59.533661 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 27 08:21:59 crc kubenswrapper[4934]: I1227 08:21:59.535656 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk"] Dec 27 08:22:00 crc kubenswrapper[4934]: I1227 08:22:00.404544 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" event={"ID":"34927e46-8eca-4b2d-ba3a-e68690b804df","Type":"ContainerStarted","Data":"9f13967c269d8163d7d57a6946462cf370ca5fb007ee80c0b25610406892356c"} Dec 27 08:22:01 crc kubenswrapper[4934]: I1227 08:22:01.421844 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" event={"ID":"34927e46-8eca-4b2d-ba3a-e68690b804df","Type":"ContainerStarted","Data":"9dac68078ec1e8d1c8a0e616f51b40143aa9f09e130f8e914e39fa5bd2cd310d"} Dec 27 08:22:01 crc kubenswrapper[4934]: I1227 08:22:01.451791 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" podStartSLOduration=2.859759026 podStartE2EDuration="3.451768073s" podCreationTimestamp="2025-12-27 08:21:58 +0000 UTC" firstStartedPulling="2025-12-27 08:21:59.533418402 +0000 UTC m=+2380.353859006" lastFinishedPulling="2025-12-27 08:22:00.125427429 +0000 UTC m=+2380.945868053" observedRunningTime="2025-12-27 08:22:01.43913834 +0000 UTC m=+2382.259578944" watchObservedRunningTime="2025-12-27 08:22:01.451768073 +0000 UTC m=+2382.272208677" Dec 27 08:22:11 crc kubenswrapper[4934]: I1227 08:22:11.468732 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:22:11 crc kubenswrapper[4934]: E1227 08:22:11.471003 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:22:23 crc kubenswrapper[4934]: I1227 08:22:23.469212 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:22:23 crc kubenswrapper[4934]: E1227 08:22:23.470663 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:22:38 crc kubenswrapper[4934]: I1227 08:22:38.468297 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:22:38 crc kubenswrapper[4934]: E1227 08:22:38.469514 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:22:52 crc kubenswrapper[4934]: I1227 08:22:52.467993 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:22:52 crc kubenswrapper[4934]: E1227 08:22:52.469065 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:22:56 crc kubenswrapper[4934]: I1227 08:22:56.215210 4934 generic.go:334] "Generic (PLEG): container finished" podID="34927e46-8eca-4b2d-ba3a-e68690b804df" containerID="9dac68078ec1e8d1c8a0e616f51b40143aa9f09e130f8e914e39fa5bd2cd310d" exitCode=0 Dec 27 08:22:56 crc kubenswrapper[4934]: I1227 08:22:56.215285 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" event={"ID":"34927e46-8eca-4b2d-ba3a-e68690b804df","Type":"ContainerDied","Data":"9dac68078ec1e8d1c8a0e616f51b40143aa9f09e130f8e914e39fa5bd2cd310d"} Dec 27 08:22:57 crc kubenswrapper[4934]: I1227 08:22:57.807137 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:22:57 crc kubenswrapper[4934]: I1227 08:22:57.864540 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2wh2\" (UniqueName: \"kubernetes.io/projected/34927e46-8eca-4b2d-ba3a-e68690b804df-kube-api-access-c2wh2\") pod \"34927e46-8eca-4b2d-ba3a-e68690b804df\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " Dec 27 08:22:57 crc kubenswrapper[4934]: I1227 08:22:57.864627 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-nova-metadata-neutron-config-0\") pod \"34927e46-8eca-4b2d-ba3a-e68690b804df\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " Dec 27 08:22:57 crc kubenswrapper[4934]: I1227 08:22:57.864682 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-neutron-metadata-combined-ca-bundle\") pod \"34927e46-8eca-4b2d-ba3a-e68690b804df\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " Dec 27 08:22:57 crc kubenswrapper[4934]: I1227 08:22:57.864830 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-inventory\") pod \"34927e46-8eca-4b2d-ba3a-e68690b804df\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " Dec 27 08:22:57 crc kubenswrapper[4934]: I1227 08:22:57.864906 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-neutron-ovn-metadata-agent-neutron-config-0\") pod \"34927e46-8eca-4b2d-ba3a-e68690b804df\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " Dec 27 08:22:57 crc kubenswrapper[4934]: I1227 08:22:57.864998 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-ssh-key\") pod \"34927e46-8eca-4b2d-ba3a-e68690b804df\" (UID: \"34927e46-8eca-4b2d-ba3a-e68690b804df\") " Dec 27 08:22:57 crc kubenswrapper[4934]: I1227 08:22:57.873606 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "34927e46-8eca-4b2d-ba3a-e68690b804df" (UID: "34927e46-8eca-4b2d-ba3a-e68690b804df"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:22:57 crc kubenswrapper[4934]: I1227 08:22:57.873757 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34927e46-8eca-4b2d-ba3a-e68690b804df-kube-api-access-c2wh2" (OuterVolumeSpecName: "kube-api-access-c2wh2") pod "34927e46-8eca-4b2d-ba3a-e68690b804df" (UID: "34927e46-8eca-4b2d-ba3a-e68690b804df"). InnerVolumeSpecName "kube-api-access-c2wh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:22:57 crc kubenswrapper[4934]: I1227 08:22:57.905232 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-inventory" (OuterVolumeSpecName: "inventory") pod "34927e46-8eca-4b2d-ba3a-e68690b804df" (UID: "34927e46-8eca-4b2d-ba3a-e68690b804df"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:22:57 crc kubenswrapper[4934]: I1227 08:22:57.919321 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "34927e46-8eca-4b2d-ba3a-e68690b804df" (UID: "34927e46-8eca-4b2d-ba3a-e68690b804df"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:22:57 crc kubenswrapper[4934]: I1227 08:22:57.922315 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "34927e46-8eca-4b2d-ba3a-e68690b804df" (UID: "34927e46-8eca-4b2d-ba3a-e68690b804df"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:22:57 crc kubenswrapper[4934]: I1227 08:22:57.925677 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "34927e46-8eca-4b2d-ba3a-e68690b804df" (UID: "34927e46-8eca-4b2d-ba3a-e68690b804df"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:22:57 crc kubenswrapper[4934]: I1227 08:22:57.968245 4934 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:22:57 crc kubenswrapper[4934]: I1227 08:22:57.968300 4934 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 27 08:22:57 crc kubenswrapper[4934]: I1227 08:22:57.968314 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2wh2\" (UniqueName: \"kubernetes.io/projected/34927e46-8eca-4b2d-ba3a-e68690b804df-kube-api-access-c2wh2\") on node \"crc\" DevicePath \"\"" Dec 27 08:22:57 crc kubenswrapper[4934]: I1227 08:22:57.968327 4934 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:22:57 crc kubenswrapper[4934]: I1227 08:22:57.968340 4934 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:22:57 crc kubenswrapper[4934]: I1227 08:22:57.968354 4934 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/34927e46-8eca-4b2d-ba3a-e68690b804df-inventory\") on node \"crc\" DevicePath \"\"" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.238518 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" event={"ID":"34927e46-8eca-4b2d-ba3a-e68690b804df","Type":"ContainerDied","Data":"9f13967c269d8163d7d57a6946462cf370ca5fb007ee80c0b25610406892356c"} Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.238555 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f13967c269d8163d7d57a6946462cf370ca5fb007ee80c0b25610406892356c" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.238608 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.416148 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8"] Dec 27 08:22:58 crc kubenswrapper[4934]: E1227 08:22:58.416916 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34927e46-8eca-4b2d-ba3a-e68690b804df" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.416939 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="34927e46-8eca-4b2d-ba3a-e68690b804df" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.417221 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="34927e46-8eca-4b2d-ba3a-e68690b804df" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.417997 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.429595 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.430573 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.430775 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.430806 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.439456 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8"] Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.444947 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-98ddn" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.485042 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8\" (UID: \"c9000388-2dbb-445d-91ed-06b1952ba4bc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.485147 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8\" (UID: \"c9000388-2dbb-445d-91ed-06b1952ba4bc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.485171 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8\" (UID: \"c9000388-2dbb-445d-91ed-06b1952ba4bc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.485287 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-782kn\" (UniqueName: \"kubernetes.io/projected/c9000388-2dbb-445d-91ed-06b1952ba4bc-kube-api-access-782kn\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8\" (UID: \"c9000388-2dbb-445d-91ed-06b1952ba4bc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.485316 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8\" (UID: \"c9000388-2dbb-445d-91ed-06b1952ba4bc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.587763 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8\" (UID: \"c9000388-2dbb-445d-91ed-06b1952ba4bc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.587828 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8\" (UID: \"c9000388-2dbb-445d-91ed-06b1952ba4bc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.587858 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8\" (UID: \"c9000388-2dbb-445d-91ed-06b1952ba4bc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.587932 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-782kn\" (UniqueName: \"kubernetes.io/projected/c9000388-2dbb-445d-91ed-06b1952ba4bc-kube-api-access-782kn\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8\" (UID: \"c9000388-2dbb-445d-91ed-06b1952ba4bc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.587954 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8\" (UID: \"c9000388-2dbb-445d-91ed-06b1952ba4bc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.592777 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8\" (UID: \"c9000388-2dbb-445d-91ed-06b1952ba4bc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.593020 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8\" (UID: \"c9000388-2dbb-445d-91ed-06b1952ba4bc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.594817 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8\" (UID: \"c9000388-2dbb-445d-91ed-06b1952ba4bc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.600919 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8\" (UID: \"c9000388-2dbb-445d-91ed-06b1952ba4bc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.604093 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-782kn\" (UniqueName: \"kubernetes.io/projected/c9000388-2dbb-445d-91ed-06b1952ba4bc-kube-api-access-782kn\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8\" (UID: \"c9000388-2dbb-445d-91ed-06b1952ba4bc\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" Dec 27 08:22:58 crc kubenswrapper[4934]: I1227 08:22:58.737473 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" Dec 27 08:22:59 crc kubenswrapper[4934]: I1227 08:22:59.384455 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8"] Dec 27 08:23:00 crc kubenswrapper[4934]: I1227 08:23:00.264006 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" event={"ID":"c9000388-2dbb-445d-91ed-06b1952ba4bc","Type":"ContainerStarted","Data":"5479098cdac4daa1db99b17e111d23a513317fd7fe90f445c6932d18ccc35f7d"} Dec 27 08:23:00 crc kubenswrapper[4934]: I1227 08:23:00.264578 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" event={"ID":"c9000388-2dbb-445d-91ed-06b1952ba4bc","Type":"ContainerStarted","Data":"c348f41cf6834a5dee95bf750dd40b1f8fe3663ca424afe26874dbe62d57b812"} Dec 27 08:23:00 crc kubenswrapper[4934]: I1227 08:23:00.288301 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" podStartSLOduration=1.8290295269999999 podStartE2EDuration="2.288280907s" podCreationTimestamp="2025-12-27 08:22:58 +0000 UTC" firstStartedPulling="2025-12-27 08:22:59.386928612 +0000 UTC m=+2440.207369246" lastFinishedPulling="2025-12-27 08:22:59.846179982 +0000 UTC m=+2440.666620626" observedRunningTime="2025-12-27 08:23:00.282720999 +0000 UTC m=+2441.103161623" watchObservedRunningTime="2025-12-27 08:23:00.288280907 +0000 UTC m=+2441.108721501" Dec 27 08:23:00 crc kubenswrapper[4934]: I1227 08:23:00.311022 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h84mn"] Dec 27 08:23:00 crc kubenswrapper[4934]: I1227 08:23:00.313378 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h84mn" Dec 27 08:23:00 crc kubenswrapper[4934]: I1227 08:23:00.323031 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h84mn"] Dec 27 08:23:00 crc kubenswrapper[4934]: I1227 08:23:00.437508 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq4gs\" (UniqueName: \"kubernetes.io/projected/24d10ddc-1a1f-4d09-a027-aed2169453a5-kube-api-access-vq4gs\") pod \"community-operators-h84mn\" (UID: \"24d10ddc-1a1f-4d09-a027-aed2169453a5\") " pod="openshift-marketplace/community-operators-h84mn" Dec 27 08:23:00 crc kubenswrapper[4934]: I1227 08:23:00.437554 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24d10ddc-1a1f-4d09-a027-aed2169453a5-utilities\") pod \"community-operators-h84mn\" (UID: \"24d10ddc-1a1f-4d09-a027-aed2169453a5\") " pod="openshift-marketplace/community-operators-h84mn" Dec 27 08:23:00 crc kubenswrapper[4934]: I1227 08:23:00.437588 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24d10ddc-1a1f-4d09-a027-aed2169453a5-catalog-content\") pod \"community-operators-h84mn\" (UID: \"24d10ddc-1a1f-4d09-a027-aed2169453a5\") " pod="openshift-marketplace/community-operators-h84mn" Dec 27 08:23:00 crc kubenswrapper[4934]: I1227 08:23:00.540206 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq4gs\" (UniqueName: \"kubernetes.io/projected/24d10ddc-1a1f-4d09-a027-aed2169453a5-kube-api-access-vq4gs\") pod \"community-operators-h84mn\" (UID: \"24d10ddc-1a1f-4d09-a027-aed2169453a5\") " pod="openshift-marketplace/community-operators-h84mn" Dec 27 08:23:00 crc kubenswrapper[4934]: I1227 08:23:00.540250 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24d10ddc-1a1f-4d09-a027-aed2169453a5-utilities\") pod \"community-operators-h84mn\" (UID: \"24d10ddc-1a1f-4d09-a027-aed2169453a5\") " pod="openshift-marketplace/community-operators-h84mn" Dec 27 08:23:00 crc kubenswrapper[4934]: I1227 08:23:00.540278 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24d10ddc-1a1f-4d09-a027-aed2169453a5-catalog-content\") pod \"community-operators-h84mn\" (UID: \"24d10ddc-1a1f-4d09-a027-aed2169453a5\") " pod="openshift-marketplace/community-operators-h84mn" Dec 27 08:23:00 crc kubenswrapper[4934]: I1227 08:23:00.541502 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24d10ddc-1a1f-4d09-a027-aed2169453a5-utilities\") pod \"community-operators-h84mn\" (UID: \"24d10ddc-1a1f-4d09-a027-aed2169453a5\") " pod="openshift-marketplace/community-operators-h84mn" Dec 27 08:23:00 crc kubenswrapper[4934]: I1227 08:23:00.541587 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24d10ddc-1a1f-4d09-a027-aed2169453a5-catalog-content\") pod \"community-operators-h84mn\" (UID: \"24d10ddc-1a1f-4d09-a027-aed2169453a5\") " pod="openshift-marketplace/community-operators-h84mn" Dec 27 08:23:00 crc kubenswrapper[4934]: I1227 08:23:00.565620 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq4gs\" (UniqueName: \"kubernetes.io/projected/24d10ddc-1a1f-4d09-a027-aed2169453a5-kube-api-access-vq4gs\") pod \"community-operators-h84mn\" (UID: \"24d10ddc-1a1f-4d09-a027-aed2169453a5\") " pod="openshift-marketplace/community-operators-h84mn" Dec 27 08:23:00 crc kubenswrapper[4934]: I1227 08:23:00.633183 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h84mn" Dec 27 08:23:01 crc kubenswrapper[4934]: W1227 08:23:01.253662 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24d10ddc_1a1f_4d09_a027_aed2169453a5.slice/crio-42c7d89091ec642bf504fcd4dd5a23dc8d473a7a0cbdd66a9a40d63b52be50a5 WatchSource:0}: Error finding container 42c7d89091ec642bf504fcd4dd5a23dc8d473a7a0cbdd66a9a40d63b52be50a5: Status 404 returned error can't find the container with id 42c7d89091ec642bf504fcd4dd5a23dc8d473a7a0cbdd66a9a40d63b52be50a5 Dec 27 08:23:01 crc kubenswrapper[4934]: I1227 08:23:01.258216 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h84mn"] Dec 27 08:23:01 crc kubenswrapper[4934]: I1227 08:23:01.275317 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h84mn" event={"ID":"24d10ddc-1a1f-4d09-a027-aed2169453a5","Type":"ContainerStarted","Data":"42c7d89091ec642bf504fcd4dd5a23dc8d473a7a0cbdd66a9a40d63b52be50a5"} Dec 27 08:23:02 crc kubenswrapper[4934]: I1227 08:23:02.292584 4934 generic.go:334] "Generic (PLEG): container finished" podID="24d10ddc-1a1f-4d09-a027-aed2169453a5" containerID="5bfa24035a9fb9299366b490bf256f419f1293a0961367cf6b0a044812305d40" exitCode=0 Dec 27 08:23:02 crc kubenswrapper[4934]: I1227 08:23:02.292650 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h84mn" event={"ID":"24d10ddc-1a1f-4d09-a027-aed2169453a5","Type":"ContainerDied","Data":"5bfa24035a9fb9299366b490bf256f419f1293a0961367cf6b0a044812305d40"} Dec 27 08:23:03 crc kubenswrapper[4934]: I1227 08:23:03.306545 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h84mn" event={"ID":"24d10ddc-1a1f-4d09-a027-aed2169453a5","Type":"ContainerStarted","Data":"24ef2a8855360bed2c6c362212156c1698a28cfef8446ff41bae4d81a99f4802"} Dec 27 08:23:04 crc kubenswrapper[4934]: I1227 08:23:04.322716 4934 generic.go:334] "Generic (PLEG): container finished" podID="24d10ddc-1a1f-4d09-a027-aed2169453a5" containerID="24ef2a8855360bed2c6c362212156c1698a28cfef8446ff41bae4d81a99f4802" exitCode=0 Dec 27 08:23:04 crc kubenswrapper[4934]: I1227 08:23:04.322782 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h84mn" event={"ID":"24d10ddc-1a1f-4d09-a027-aed2169453a5","Type":"ContainerDied","Data":"24ef2a8855360bed2c6c362212156c1698a28cfef8446ff41bae4d81a99f4802"} Dec 27 08:23:04 crc kubenswrapper[4934]: I1227 08:23:04.468513 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:23:04 crc kubenswrapper[4934]: E1227 08:23:04.468937 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:23:05 crc kubenswrapper[4934]: I1227 08:23:05.341823 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h84mn" event={"ID":"24d10ddc-1a1f-4d09-a027-aed2169453a5","Type":"ContainerStarted","Data":"d9d2066d67e809102cd6c447037277ef6cc1a647153fe97e17ed03575b84fa89"} Dec 27 08:23:05 crc kubenswrapper[4934]: I1227 08:23:05.363769 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h84mn" podStartSLOduration=2.899454337 podStartE2EDuration="5.363744602s" podCreationTimestamp="2025-12-27 08:23:00 +0000 UTC" firstStartedPulling="2025-12-27 08:23:02.29649449 +0000 UTC m=+2443.116935114" lastFinishedPulling="2025-12-27 08:23:04.760784785 +0000 UTC m=+2445.581225379" observedRunningTime="2025-12-27 08:23:05.360990514 +0000 UTC m=+2446.181431178" watchObservedRunningTime="2025-12-27 08:23:05.363744602 +0000 UTC m=+2446.184185196" Dec 27 08:23:10 crc kubenswrapper[4934]: I1227 08:23:10.633732 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h84mn" Dec 27 08:23:10 crc kubenswrapper[4934]: I1227 08:23:10.633797 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h84mn" Dec 27 08:23:10 crc kubenswrapper[4934]: I1227 08:23:10.688156 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h84mn" Dec 27 08:23:11 crc kubenswrapper[4934]: I1227 08:23:11.501885 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h84mn" Dec 27 08:23:11 crc kubenswrapper[4934]: I1227 08:23:11.566216 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h84mn"] Dec 27 08:23:13 crc kubenswrapper[4934]: I1227 08:23:13.441134 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-h84mn" podUID="24d10ddc-1a1f-4d09-a027-aed2169453a5" containerName="registry-server" containerID="cri-o://d9d2066d67e809102cd6c447037277ef6cc1a647153fe97e17ed03575b84fa89" gracePeriod=2 Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.055800 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h84mn" Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.119813 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24d10ddc-1a1f-4d09-a027-aed2169453a5-utilities\") pod \"24d10ddc-1a1f-4d09-a027-aed2169453a5\" (UID: \"24d10ddc-1a1f-4d09-a027-aed2169453a5\") " Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.120111 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vq4gs\" (UniqueName: \"kubernetes.io/projected/24d10ddc-1a1f-4d09-a027-aed2169453a5-kube-api-access-vq4gs\") pod \"24d10ddc-1a1f-4d09-a027-aed2169453a5\" (UID: \"24d10ddc-1a1f-4d09-a027-aed2169453a5\") " Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.120309 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24d10ddc-1a1f-4d09-a027-aed2169453a5-catalog-content\") pod \"24d10ddc-1a1f-4d09-a027-aed2169453a5\" (UID: \"24d10ddc-1a1f-4d09-a027-aed2169453a5\") " Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.120787 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24d10ddc-1a1f-4d09-a027-aed2169453a5-utilities" (OuterVolumeSpecName: "utilities") pod "24d10ddc-1a1f-4d09-a027-aed2169453a5" (UID: "24d10ddc-1a1f-4d09-a027-aed2169453a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.121008 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24d10ddc-1a1f-4d09-a027-aed2169453a5-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.140408 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24d10ddc-1a1f-4d09-a027-aed2169453a5-kube-api-access-vq4gs" (OuterVolumeSpecName: "kube-api-access-vq4gs") pod "24d10ddc-1a1f-4d09-a027-aed2169453a5" (UID: "24d10ddc-1a1f-4d09-a027-aed2169453a5"). InnerVolumeSpecName "kube-api-access-vq4gs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.222783 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vq4gs\" (UniqueName: \"kubernetes.io/projected/24d10ddc-1a1f-4d09-a027-aed2169453a5-kube-api-access-vq4gs\") on node \"crc\" DevicePath \"\"" Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.267481 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24d10ddc-1a1f-4d09-a027-aed2169453a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "24d10ddc-1a1f-4d09-a027-aed2169453a5" (UID: "24d10ddc-1a1f-4d09-a027-aed2169453a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.325542 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24d10ddc-1a1f-4d09-a027-aed2169453a5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.456643 4934 generic.go:334] "Generic (PLEG): container finished" podID="24d10ddc-1a1f-4d09-a027-aed2169453a5" containerID="d9d2066d67e809102cd6c447037277ef6cc1a647153fe97e17ed03575b84fa89" exitCode=0 Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.456716 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h84mn" event={"ID":"24d10ddc-1a1f-4d09-a027-aed2169453a5","Type":"ContainerDied","Data":"d9d2066d67e809102cd6c447037277ef6cc1a647153fe97e17ed03575b84fa89"} Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.456738 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h84mn" Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.456761 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h84mn" event={"ID":"24d10ddc-1a1f-4d09-a027-aed2169453a5","Type":"ContainerDied","Data":"42c7d89091ec642bf504fcd4dd5a23dc8d473a7a0cbdd66a9a40d63b52be50a5"} Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.456792 4934 scope.go:117] "RemoveContainer" containerID="d9d2066d67e809102cd6c447037277ef6cc1a647153fe97e17ed03575b84fa89" Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.488222 4934 scope.go:117] "RemoveContainer" containerID="24ef2a8855360bed2c6c362212156c1698a28cfef8446ff41bae4d81a99f4802" Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.527138 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h84mn"] Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.532152 4934 scope.go:117] "RemoveContainer" containerID="5bfa24035a9fb9299366b490bf256f419f1293a0961367cf6b0a044812305d40" Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.536382 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-h84mn"] Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.624613 4934 scope.go:117] "RemoveContainer" containerID="d9d2066d67e809102cd6c447037277ef6cc1a647153fe97e17ed03575b84fa89" Dec 27 08:23:14 crc kubenswrapper[4934]: E1227 08:23:14.625186 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9d2066d67e809102cd6c447037277ef6cc1a647153fe97e17ed03575b84fa89\": container with ID starting with d9d2066d67e809102cd6c447037277ef6cc1a647153fe97e17ed03575b84fa89 not found: ID does not exist" containerID="d9d2066d67e809102cd6c447037277ef6cc1a647153fe97e17ed03575b84fa89" Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.625232 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9d2066d67e809102cd6c447037277ef6cc1a647153fe97e17ed03575b84fa89"} err="failed to get container status \"d9d2066d67e809102cd6c447037277ef6cc1a647153fe97e17ed03575b84fa89\": rpc error: code = NotFound desc = could not find container \"d9d2066d67e809102cd6c447037277ef6cc1a647153fe97e17ed03575b84fa89\": container with ID starting with d9d2066d67e809102cd6c447037277ef6cc1a647153fe97e17ed03575b84fa89 not found: ID does not exist" Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.625264 4934 scope.go:117] "RemoveContainer" containerID="24ef2a8855360bed2c6c362212156c1698a28cfef8446ff41bae4d81a99f4802" Dec 27 08:23:14 crc kubenswrapper[4934]: E1227 08:23:14.625788 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24ef2a8855360bed2c6c362212156c1698a28cfef8446ff41bae4d81a99f4802\": container with ID starting with 24ef2a8855360bed2c6c362212156c1698a28cfef8446ff41bae4d81a99f4802 not found: ID does not exist" containerID="24ef2a8855360bed2c6c362212156c1698a28cfef8446ff41bae4d81a99f4802" Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.625822 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24ef2a8855360bed2c6c362212156c1698a28cfef8446ff41bae4d81a99f4802"} err="failed to get container status \"24ef2a8855360bed2c6c362212156c1698a28cfef8446ff41bae4d81a99f4802\": rpc error: code = NotFound desc = could not find container \"24ef2a8855360bed2c6c362212156c1698a28cfef8446ff41bae4d81a99f4802\": container with ID starting with 24ef2a8855360bed2c6c362212156c1698a28cfef8446ff41bae4d81a99f4802 not found: ID does not exist" Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.625847 4934 scope.go:117] "RemoveContainer" containerID="5bfa24035a9fb9299366b490bf256f419f1293a0961367cf6b0a044812305d40" Dec 27 08:23:14 crc kubenswrapper[4934]: E1227 08:23:14.626261 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bfa24035a9fb9299366b490bf256f419f1293a0961367cf6b0a044812305d40\": container with ID starting with 5bfa24035a9fb9299366b490bf256f419f1293a0961367cf6b0a044812305d40 not found: ID does not exist" containerID="5bfa24035a9fb9299366b490bf256f419f1293a0961367cf6b0a044812305d40" Dec 27 08:23:14 crc kubenswrapper[4934]: I1227 08:23:14.626292 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bfa24035a9fb9299366b490bf256f419f1293a0961367cf6b0a044812305d40"} err="failed to get container status \"5bfa24035a9fb9299366b490bf256f419f1293a0961367cf6b0a044812305d40\": rpc error: code = NotFound desc = could not find container \"5bfa24035a9fb9299366b490bf256f419f1293a0961367cf6b0a044812305d40\": container with ID starting with 5bfa24035a9fb9299366b490bf256f419f1293a0961367cf6b0a044812305d40 not found: ID does not exist" Dec 27 08:23:15 crc kubenswrapper[4934]: I1227 08:23:15.489012 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24d10ddc-1a1f-4d09-a027-aed2169453a5" path="/var/lib/kubelet/pods/24d10ddc-1a1f-4d09-a027-aed2169453a5/volumes" Dec 27 08:23:19 crc kubenswrapper[4934]: I1227 08:23:19.483145 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:23:19 crc kubenswrapper[4934]: E1227 08:23:19.483941 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:23:30 crc kubenswrapper[4934]: I1227 08:23:30.468157 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:23:30 crc kubenswrapper[4934]: E1227 08:23:30.469403 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:23:41 crc kubenswrapper[4934]: I1227 08:23:41.471782 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:23:41 crc kubenswrapper[4934]: E1227 08:23:41.472710 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:23:53 crc kubenswrapper[4934]: I1227 08:23:53.468856 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:23:53 crc kubenswrapper[4934]: E1227 08:23:53.470234 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:24:05 crc kubenswrapper[4934]: I1227 08:24:05.468034 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:24:05 crc kubenswrapper[4934]: E1227 08:24:05.469602 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:24:20 crc kubenswrapper[4934]: I1227 08:24:20.468576 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:24:20 crc kubenswrapper[4934]: E1227 08:24:20.469198 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:24:31 crc kubenswrapper[4934]: I1227 08:24:31.468316 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:24:31 crc kubenswrapper[4934]: E1227 08:24:31.469259 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:24:46 crc kubenswrapper[4934]: I1227 08:24:46.467744 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:24:47 crc kubenswrapper[4934]: I1227 08:24:47.446631 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"0b5192c2223a1f8a1927e3908026b1994ec18d42dcd71c448b0446cd8d833be2"} Dec 27 08:27:15 crc kubenswrapper[4934]: I1227 08:27:15.330416 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:27:15 crc kubenswrapper[4934]: I1227 08:27:15.331074 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:27:45 crc kubenswrapper[4934]: I1227 08:27:45.330415 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:27:45 crc kubenswrapper[4934]: I1227 08:27:45.331056 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:27:46 crc kubenswrapper[4934]: I1227 08:27:46.960813 4934 generic.go:334] "Generic (PLEG): container finished" podID="c9000388-2dbb-445d-91ed-06b1952ba4bc" containerID="5479098cdac4daa1db99b17e111d23a513317fd7fe90f445c6932d18ccc35f7d" exitCode=0 Dec 27 08:27:46 crc kubenswrapper[4934]: I1227 08:27:46.960895 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" event={"ID":"c9000388-2dbb-445d-91ed-06b1952ba4bc","Type":"ContainerDied","Data":"5479098cdac4daa1db99b17e111d23a513317fd7fe90f445c6932d18ccc35f7d"} Dec 27 08:27:48 crc kubenswrapper[4934]: I1227 08:27:48.494618 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" Dec 27 08:27:48 crc kubenswrapper[4934]: I1227 08:27:48.642191 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-libvirt-secret-0\") pod \"c9000388-2dbb-445d-91ed-06b1952ba4bc\" (UID: \"c9000388-2dbb-445d-91ed-06b1952ba4bc\") " Dec 27 08:27:48 crc kubenswrapper[4934]: I1227 08:27:48.642257 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-782kn\" (UniqueName: \"kubernetes.io/projected/c9000388-2dbb-445d-91ed-06b1952ba4bc-kube-api-access-782kn\") pod \"c9000388-2dbb-445d-91ed-06b1952ba4bc\" (UID: \"c9000388-2dbb-445d-91ed-06b1952ba4bc\") " Dec 27 08:27:48 crc kubenswrapper[4934]: I1227 08:27:48.642314 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-ssh-key\") pod \"c9000388-2dbb-445d-91ed-06b1952ba4bc\" (UID: \"c9000388-2dbb-445d-91ed-06b1952ba4bc\") " Dec 27 08:27:48 crc kubenswrapper[4934]: I1227 08:27:48.642476 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-libvirt-combined-ca-bundle\") pod \"c9000388-2dbb-445d-91ed-06b1952ba4bc\" (UID: \"c9000388-2dbb-445d-91ed-06b1952ba4bc\") " Dec 27 08:27:48 crc kubenswrapper[4934]: I1227 08:27:48.642586 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-inventory\") pod \"c9000388-2dbb-445d-91ed-06b1952ba4bc\" (UID: \"c9000388-2dbb-445d-91ed-06b1952ba4bc\") " Dec 27 08:27:48 crc kubenswrapper[4934]: I1227 08:27:48.648252 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "c9000388-2dbb-445d-91ed-06b1952ba4bc" (UID: "c9000388-2dbb-445d-91ed-06b1952ba4bc"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:27:48 crc kubenswrapper[4934]: I1227 08:27:48.649185 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9000388-2dbb-445d-91ed-06b1952ba4bc-kube-api-access-782kn" (OuterVolumeSpecName: "kube-api-access-782kn") pod "c9000388-2dbb-445d-91ed-06b1952ba4bc" (UID: "c9000388-2dbb-445d-91ed-06b1952ba4bc"). InnerVolumeSpecName "kube-api-access-782kn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:27:48 crc kubenswrapper[4934]: I1227 08:27:48.673256 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "c9000388-2dbb-445d-91ed-06b1952ba4bc" (UID: "c9000388-2dbb-445d-91ed-06b1952ba4bc"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:27:48 crc kubenswrapper[4934]: I1227 08:27:48.675731 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c9000388-2dbb-445d-91ed-06b1952ba4bc" (UID: "c9000388-2dbb-445d-91ed-06b1952ba4bc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:27:48 crc kubenswrapper[4934]: I1227 08:27:48.676283 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-inventory" (OuterVolumeSpecName: "inventory") pod "c9000388-2dbb-445d-91ed-06b1952ba4bc" (UID: "c9000388-2dbb-445d-91ed-06b1952ba4bc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:27:48 crc kubenswrapper[4934]: I1227 08:27:48.745473 4934 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:27:48 crc kubenswrapper[4934]: I1227 08:27:48.745509 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-782kn\" (UniqueName: \"kubernetes.io/projected/c9000388-2dbb-445d-91ed-06b1952ba4bc-kube-api-access-782kn\") on node \"crc\" DevicePath \"\"" Dec 27 08:27:48 crc kubenswrapper[4934]: I1227 08:27:48.745522 4934 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 27 08:27:48 crc kubenswrapper[4934]: I1227 08:27:48.745532 4934 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:27:48 crc kubenswrapper[4934]: I1227 08:27:48.745541 4934 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9000388-2dbb-445d-91ed-06b1952ba4bc-inventory\") on node \"crc\" DevicePath \"\"" Dec 27 08:27:48 crc kubenswrapper[4934]: I1227 08:27:48.983321 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" event={"ID":"c9000388-2dbb-445d-91ed-06b1952ba4bc","Type":"ContainerDied","Data":"c348f41cf6834a5dee95bf750dd40b1f8fe3663ca424afe26874dbe62d57b812"} Dec 27 08:27:48 crc kubenswrapper[4934]: I1227 08:27:48.983668 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c348f41cf6834a5dee95bf750dd40b1f8fe3663ca424afe26874dbe62d57b812" Dec 27 08:27:48 crc kubenswrapper[4934]: I1227 08:27:48.983494 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.141554 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2"] Dec 27 08:27:49 crc kubenswrapper[4934]: E1227 08:27:49.142432 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24d10ddc-1a1f-4d09-a027-aed2169453a5" containerName="registry-server" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.142465 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="24d10ddc-1a1f-4d09-a027-aed2169453a5" containerName="registry-server" Dec 27 08:27:49 crc kubenswrapper[4934]: E1227 08:27:49.142483 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24d10ddc-1a1f-4d09-a027-aed2169453a5" containerName="extract-content" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.142496 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="24d10ddc-1a1f-4d09-a027-aed2169453a5" containerName="extract-content" Dec 27 08:27:49 crc kubenswrapper[4934]: E1227 08:27:49.142528 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9000388-2dbb-445d-91ed-06b1952ba4bc" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.142544 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9000388-2dbb-445d-91ed-06b1952ba4bc" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 27 08:27:49 crc kubenswrapper[4934]: E1227 08:27:49.142584 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24d10ddc-1a1f-4d09-a027-aed2169453a5" containerName="extract-utilities" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.142596 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="24d10ddc-1a1f-4d09-a027-aed2169453a5" containerName="extract-utilities" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.143115 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="24d10ddc-1a1f-4d09-a027-aed2169453a5" containerName="registry-server" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.143165 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9000388-2dbb-445d-91ed-06b1952ba4bc" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.144705 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.153693 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.154301 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.154433 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.154545 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.154693 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-98ddn" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.154819 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.155529 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.167801 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2"] Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.255857 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.255913 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.256215 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.256268 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.256531 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.256735 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.256997 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4v79\" (UniqueName: \"kubernetes.io/projected/2f4288f0-95f7-4b96-8066-e8b892d401ae-kube-api-access-w4v79\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.257145 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.257192 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.359039 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.359101 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.360023 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.360134 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.360247 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4v79\" (UniqueName: \"kubernetes.io/projected/2f4288f0-95f7-4b96-8066-e8b892d401ae-kube-api-access-w4v79\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.360304 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.360336 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.360378 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.360427 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.361730 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.365213 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.366444 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.366789 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.367238 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.368336 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.368751 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.380325 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.383286 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4v79\" (UniqueName: \"kubernetes.io/projected/2f4288f0-95f7-4b96-8066-e8b892d401ae-kube-api-access-w4v79\") pod \"nova-edpm-deployment-openstack-edpm-ipam-qb7m2\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:49 crc kubenswrapper[4934]: I1227 08:27:49.478467 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:27:50 crc kubenswrapper[4934]: W1227 08:27:50.081149 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f4288f0_95f7_4b96_8066_e8b892d401ae.slice/crio-ebd82b0d988be348057fe3ff273529d3991ade7822c9010d1f2e68a1bbf6c900 WatchSource:0}: Error finding container ebd82b0d988be348057fe3ff273529d3991ade7822c9010d1f2e68a1bbf6c900: Status 404 returned error can't find the container with id ebd82b0d988be348057fe3ff273529d3991ade7822c9010d1f2e68a1bbf6c900 Dec 27 08:27:50 crc kubenswrapper[4934]: I1227 08:27:50.083146 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2"] Dec 27 08:27:50 crc kubenswrapper[4934]: I1227 08:27:50.089783 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 27 08:27:51 crc kubenswrapper[4934]: I1227 08:27:51.012362 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" event={"ID":"2f4288f0-95f7-4b96-8066-e8b892d401ae","Type":"ContainerStarted","Data":"ebd82b0d988be348057fe3ff273529d3991ade7822c9010d1f2e68a1bbf6c900"} Dec 27 08:27:52 crc kubenswrapper[4934]: I1227 08:27:52.024229 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" event={"ID":"2f4288f0-95f7-4b96-8066-e8b892d401ae","Type":"ContainerStarted","Data":"f9dc6bcada6d7de5b3b06d56d9ddec2526649242476946d4d992e13681c530d6"} Dec 27 08:27:52 crc kubenswrapper[4934]: I1227 08:27:52.054130 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" podStartSLOduration=2.406066736 podStartE2EDuration="3.054110872s" podCreationTimestamp="2025-12-27 08:27:49 +0000 UTC" firstStartedPulling="2025-12-27 08:27:50.089395096 +0000 UTC m=+2730.909835700" lastFinishedPulling="2025-12-27 08:27:50.737439232 +0000 UTC m=+2731.557879836" observedRunningTime="2025-12-27 08:27:52.043722905 +0000 UTC m=+2732.864163499" watchObservedRunningTime="2025-12-27 08:27:52.054110872 +0000 UTC m=+2732.874551466" Dec 27 08:28:15 crc kubenswrapper[4934]: I1227 08:28:15.330500 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:28:15 crc kubenswrapper[4934]: I1227 08:28:15.331000 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:28:15 crc kubenswrapper[4934]: I1227 08:28:15.331050 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 08:28:15 crc kubenswrapper[4934]: I1227 08:28:15.331983 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0b5192c2223a1f8a1927e3908026b1994ec18d42dcd71c448b0446cd8d833be2"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 08:28:15 crc kubenswrapper[4934]: I1227 08:28:15.332045 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://0b5192c2223a1f8a1927e3908026b1994ec18d42dcd71c448b0446cd8d833be2" gracePeriod=600 Dec 27 08:28:16 crc kubenswrapper[4934]: I1227 08:28:16.330484 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="0b5192c2223a1f8a1927e3908026b1994ec18d42dcd71c448b0446cd8d833be2" exitCode=0 Dec 27 08:28:16 crc kubenswrapper[4934]: I1227 08:28:16.330602 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"0b5192c2223a1f8a1927e3908026b1994ec18d42dcd71c448b0446cd8d833be2"} Dec 27 08:28:16 crc kubenswrapper[4934]: I1227 08:28:16.330989 4934 scope.go:117] "RemoveContainer" containerID="93e790f2b6926b78d765a11bcdc3fbe31521108844d3219af2166d25befef6a6" Dec 27 08:28:17 crc kubenswrapper[4934]: I1227 08:28:17.346006 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b"} Dec 27 08:28:33 crc kubenswrapper[4934]: I1227 08:28:33.910973 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pzfjl"] Dec 27 08:28:33 crc kubenswrapper[4934]: I1227 08:28:33.913902 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pzfjl" Dec 27 08:28:33 crc kubenswrapper[4934]: I1227 08:28:33.952828 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a1181e-029e-4c3f-9b41-87f7500fb6b9-utilities\") pod \"certified-operators-pzfjl\" (UID: \"88a1181e-029e-4c3f-9b41-87f7500fb6b9\") " pod="openshift-marketplace/certified-operators-pzfjl" Dec 27 08:28:33 crc kubenswrapper[4934]: I1227 08:28:33.952896 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v8n8\" (UniqueName: \"kubernetes.io/projected/88a1181e-029e-4c3f-9b41-87f7500fb6b9-kube-api-access-8v8n8\") pod \"certified-operators-pzfjl\" (UID: \"88a1181e-029e-4c3f-9b41-87f7500fb6b9\") " pod="openshift-marketplace/certified-operators-pzfjl" Dec 27 08:28:33 crc kubenswrapper[4934]: I1227 08:28:33.953040 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a1181e-029e-4c3f-9b41-87f7500fb6b9-catalog-content\") pod \"certified-operators-pzfjl\" (UID: \"88a1181e-029e-4c3f-9b41-87f7500fb6b9\") " pod="openshift-marketplace/certified-operators-pzfjl" Dec 27 08:28:33 crc kubenswrapper[4934]: I1227 08:28:33.977535 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pzfjl"] Dec 27 08:28:34 crc kubenswrapper[4934]: I1227 08:28:34.055199 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a1181e-029e-4c3f-9b41-87f7500fb6b9-catalog-content\") pod \"certified-operators-pzfjl\" (UID: \"88a1181e-029e-4c3f-9b41-87f7500fb6b9\") " pod="openshift-marketplace/certified-operators-pzfjl" Dec 27 08:28:34 crc kubenswrapper[4934]: I1227 08:28:34.055378 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a1181e-029e-4c3f-9b41-87f7500fb6b9-utilities\") pod \"certified-operators-pzfjl\" (UID: \"88a1181e-029e-4c3f-9b41-87f7500fb6b9\") " pod="openshift-marketplace/certified-operators-pzfjl" Dec 27 08:28:34 crc kubenswrapper[4934]: I1227 08:28:34.055426 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v8n8\" (UniqueName: \"kubernetes.io/projected/88a1181e-029e-4c3f-9b41-87f7500fb6b9-kube-api-access-8v8n8\") pod \"certified-operators-pzfjl\" (UID: \"88a1181e-029e-4c3f-9b41-87f7500fb6b9\") " pod="openshift-marketplace/certified-operators-pzfjl" Dec 27 08:28:34 crc kubenswrapper[4934]: I1227 08:28:34.055779 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a1181e-029e-4c3f-9b41-87f7500fb6b9-catalog-content\") pod \"certified-operators-pzfjl\" (UID: \"88a1181e-029e-4c3f-9b41-87f7500fb6b9\") " pod="openshift-marketplace/certified-operators-pzfjl" Dec 27 08:28:34 crc kubenswrapper[4934]: I1227 08:28:34.055869 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a1181e-029e-4c3f-9b41-87f7500fb6b9-utilities\") pod \"certified-operators-pzfjl\" (UID: \"88a1181e-029e-4c3f-9b41-87f7500fb6b9\") " pod="openshift-marketplace/certified-operators-pzfjl" Dec 27 08:28:34 crc kubenswrapper[4934]: I1227 08:28:34.075239 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v8n8\" (UniqueName: \"kubernetes.io/projected/88a1181e-029e-4c3f-9b41-87f7500fb6b9-kube-api-access-8v8n8\") pod \"certified-operators-pzfjl\" (UID: \"88a1181e-029e-4c3f-9b41-87f7500fb6b9\") " pod="openshift-marketplace/certified-operators-pzfjl" Dec 27 08:28:34 crc kubenswrapper[4934]: I1227 08:28:34.237133 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pzfjl" Dec 27 08:28:34 crc kubenswrapper[4934]: I1227 08:28:34.837423 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pzfjl"] Dec 27 08:28:35 crc kubenswrapper[4934]: I1227 08:28:35.580671 4934 generic.go:334] "Generic (PLEG): container finished" podID="88a1181e-029e-4c3f-9b41-87f7500fb6b9" containerID="a13b4c2e3dcd9ca8871c29ca469ad825a2723eea2f4c6c841b13cc86dbae7611" exitCode=0 Dec 27 08:28:35 crc kubenswrapper[4934]: I1227 08:28:35.580737 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzfjl" event={"ID":"88a1181e-029e-4c3f-9b41-87f7500fb6b9","Type":"ContainerDied","Data":"a13b4c2e3dcd9ca8871c29ca469ad825a2723eea2f4c6c841b13cc86dbae7611"} Dec 27 08:28:35 crc kubenswrapper[4934]: I1227 08:28:35.581064 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzfjl" event={"ID":"88a1181e-029e-4c3f-9b41-87f7500fb6b9","Type":"ContainerStarted","Data":"af5605f84e9a8996b80da7f6ed3440dd159cebcb2a808873ba99aedaa23f784b"} Dec 27 08:28:37 crc kubenswrapper[4934]: I1227 08:28:37.607075 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzfjl" event={"ID":"88a1181e-029e-4c3f-9b41-87f7500fb6b9","Type":"ContainerStarted","Data":"77f5fd748da4d21e5b294de91a97165c9fd018cd107c77d73d219da5dffd7bdc"} Dec 27 08:28:38 crc kubenswrapper[4934]: I1227 08:28:38.772233 4934 generic.go:334] "Generic (PLEG): container finished" podID="88a1181e-029e-4c3f-9b41-87f7500fb6b9" containerID="77f5fd748da4d21e5b294de91a97165c9fd018cd107c77d73d219da5dffd7bdc" exitCode=0 Dec 27 08:28:38 crc kubenswrapper[4934]: I1227 08:28:38.772639 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzfjl" event={"ID":"88a1181e-029e-4c3f-9b41-87f7500fb6b9","Type":"ContainerDied","Data":"77f5fd748da4d21e5b294de91a97165c9fd018cd107c77d73d219da5dffd7bdc"} Dec 27 08:28:40 crc kubenswrapper[4934]: I1227 08:28:40.809731 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzfjl" event={"ID":"88a1181e-029e-4c3f-9b41-87f7500fb6b9","Type":"ContainerStarted","Data":"91f665085b4f3ee6d3256c2a5ada5668676cd5456cf9d04bee8265ecd74153f8"} Dec 27 08:28:40 crc kubenswrapper[4934]: I1227 08:28:40.832727 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pzfjl" podStartSLOduration=3.811275331 podStartE2EDuration="7.832701121s" podCreationTimestamp="2025-12-27 08:28:33 +0000 UTC" firstStartedPulling="2025-12-27 08:28:35.583173079 +0000 UTC m=+2776.403613683" lastFinishedPulling="2025-12-27 08:28:39.604598879 +0000 UTC m=+2780.425039473" observedRunningTime="2025-12-27 08:28:40.828450005 +0000 UTC m=+2781.648890599" watchObservedRunningTime="2025-12-27 08:28:40.832701121 +0000 UTC m=+2781.653141705" Dec 27 08:28:44 crc kubenswrapper[4934]: I1227 08:28:44.237829 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pzfjl" Dec 27 08:28:44 crc kubenswrapper[4934]: I1227 08:28:44.238503 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pzfjl" Dec 27 08:28:45 crc kubenswrapper[4934]: I1227 08:28:45.311823 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-pzfjl" podUID="88a1181e-029e-4c3f-9b41-87f7500fb6b9" containerName="registry-server" probeResult="failure" output=< Dec 27 08:28:45 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 08:28:45 crc kubenswrapper[4934]: > Dec 27 08:28:54 crc kubenswrapper[4934]: I1227 08:28:54.336104 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pzfjl" Dec 27 08:28:54 crc kubenswrapper[4934]: I1227 08:28:54.411457 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pzfjl" Dec 27 08:28:54 crc kubenswrapper[4934]: I1227 08:28:54.577535 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pzfjl"] Dec 27 08:28:56 crc kubenswrapper[4934]: I1227 08:28:56.008893 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pzfjl" podUID="88a1181e-029e-4c3f-9b41-87f7500fb6b9" containerName="registry-server" containerID="cri-o://91f665085b4f3ee6d3256c2a5ada5668676cd5456cf9d04bee8265ecd74153f8" gracePeriod=2 Dec 27 08:28:56 crc kubenswrapper[4934]: I1227 08:28:56.684554 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pzfjl" Dec 27 08:28:56 crc kubenswrapper[4934]: I1227 08:28:56.776613 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a1181e-029e-4c3f-9b41-87f7500fb6b9-utilities\") pod \"88a1181e-029e-4c3f-9b41-87f7500fb6b9\" (UID: \"88a1181e-029e-4c3f-9b41-87f7500fb6b9\") " Dec 27 08:28:56 crc kubenswrapper[4934]: I1227 08:28:56.776767 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a1181e-029e-4c3f-9b41-87f7500fb6b9-catalog-content\") pod \"88a1181e-029e-4c3f-9b41-87f7500fb6b9\" (UID: \"88a1181e-029e-4c3f-9b41-87f7500fb6b9\") " Dec 27 08:28:56 crc kubenswrapper[4934]: I1227 08:28:56.776801 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8v8n8\" (UniqueName: \"kubernetes.io/projected/88a1181e-029e-4c3f-9b41-87f7500fb6b9-kube-api-access-8v8n8\") pod \"88a1181e-029e-4c3f-9b41-87f7500fb6b9\" (UID: \"88a1181e-029e-4c3f-9b41-87f7500fb6b9\") " Dec 27 08:28:56 crc kubenswrapper[4934]: I1227 08:28:56.777591 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88a1181e-029e-4c3f-9b41-87f7500fb6b9-utilities" (OuterVolumeSpecName: "utilities") pod "88a1181e-029e-4c3f-9b41-87f7500fb6b9" (UID: "88a1181e-029e-4c3f-9b41-87f7500fb6b9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:28:56 crc kubenswrapper[4934]: I1227 08:28:56.778244 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a1181e-029e-4c3f-9b41-87f7500fb6b9-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 08:28:56 crc kubenswrapper[4934]: I1227 08:28:56.783974 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88a1181e-029e-4c3f-9b41-87f7500fb6b9-kube-api-access-8v8n8" (OuterVolumeSpecName: "kube-api-access-8v8n8") pod "88a1181e-029e-4c3f-9b41-87f7500fb6b9" (UID: "88a1181e-029e-4c3f-9b41-87f7500fb6b9"). InnerVolumeSpecName "kube-api-access-8v8n8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:28:56 crc kubenswrapper[4934]: I1227 08:28:56.840542 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88a1181e-029e-4c3f-9b41-87f7500fb6b9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88a1181e-029e-4c3f-9b41-87f7500fb6b9" (UID: "88a1181e-029e-4c3f-9b41-87f7500fb6b9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:28:56 crc kubenswrapper[4934]: I1227 08:28:56.880814 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a1181e-029e-4c3f-9b41-87f7500fb6b9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 08:28:56 crc kubenswrapper[4934]: I1227 08:28:56.880854 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8v8n8\" (UniqueName: \"kubernetes.io/projected/88a1181e-029e-4c3f-9b41-87f7500fb6b9-kube-api-access-8v8n8\") on node \"crc\" DevicePath \"\"" Dec 27 08:28:57 crc kubenswrapper[4934]: I1227 08:28:57.027568 4934 generic.go:334] "Generic (PLEG): container finished" podID="88a1181e-029e-4c3f-9b41-87f7500fb6b9" containerID="91f665085b4f3ee6d3256c2a5ada5668676cd5456cf9d04bee8265ecd74153f8" exitCode=0 Dec 27 08:28:57 crc kubenswrapper[4934]: I1227 08:28:57.027639 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pzfjl" Dec 27 08:28:57 crc kubenswrapper[4934]: I1227 08:28:57.027651 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzfjl" event={"ID":"88a1181e-029e-4c3f-9b41-87f7500fb6b9","Type":"ContainerDied","Data":"91f665085b4f3ee6d3256c2a5ada5668676cd5456cf9d04bee8265ecd74153f8"} Dec 27 08:28:57 crc kubenswrapper[4934]: I1227 08:28:57.027976 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pzfjl" event={"ID":"88a1181e-029e-4c3f-9b41-87f7500fb6b9","Type":"ContainerDied","Data":"af5605f84e9a8996b80da7f6ed3440dd159cebcb2a808873ba99aedaa23f784b"} Dec 27 08:28:57 crc kubenswrapper[4934]: I1227 08:28:57.028000 4934 scope.go:117] "RemoveContainer" containerID="91f665085b4f3ee6d3256c2a5ada5668676cd5456cf9d04bee8265ecd74153f8" Dec 27 08:28:57 crc kubenswrapper[4934]: I1227 08:28:57.066164 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pzfjl"] Dec 27 08:28:57 crc kubenswrapper[4934]: I1227 08:28:57.071973 4934 scope.go:117] "RemoveContainer" containerID="77f5fd748da4d21e5b294de91a97165c9fd018cd107c77d73d219da5dffd7bdc" Dec 27 08:28:57 crc kubenswrapper[4934]: I1227 08:28:57.078602 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pzfjl"] Dec 27 08:28:57 crc kubenswrapper[4934]: I1227 08:28:57.105898 4934 scope.go:117] "RemoveContainer" containerID="a13b4c2e3dcd9ca8871c29ca469ad825a2723eea2f4c6c841b13cc86dbae7611" Dec 27 08:28:57 crc kubenswrapper[4934]: I1227 08:28:57.155697 4934 scope.go:117] "RemoveContainer" containerID="91f665085b4f3ee6d3256c2a5ada5668676cd5456cf9d04bee8265ecd74153f8" Dec 27 08:28:57 crc kubenswrapper[4934]: E1227 08:28:57.156345 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91f665085b4f3ee6d3256c2a5ada5668676cd5456cf9d04bee8265ecd74153f8\": container with ID starting with 91f665085b4f3ee6d3256c2a5ada5668676cd5456cf9d04bee8265ecd74153f8 not found: ID does not exist" containerID="91f665085b4f3ee6d3256c2a5ada5668676cd5456cf9d04bee8265ecd74153f8" Dec 27 08:28:57 crc kubenswrapper[4934]: I1227 08:28:57.156397 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91f665085b4f3ee6d3256c2a5ada5668676cd5456cf9d04bee8265ecd74153f8"} err="failed to get container status \"91f665085b4f3ee6d3256c2a5ada5668676cd5456cf9d04bee8265ecd74153f8\": rpc error: code = NotFound desc = could not find container \"91f665085b4f3ee6d3256c2a5ada5668676cd5456cf9d04bee8265ecd74153f8\": container with ID starting with 91f665085b4f3ee6d3256c2a5ada5668676cd5456cf9d04bee8265ecd74153f8 not found: ID does not exist" Dec 27 08:28:57 crc kubenswrapper[4934]: I1227 08:28:57.156429 4934 scope.go:117] "RemoveContainer" containerID="77f5fd748da4d21e5b294de91a97165c9fd018cd107c77d73d219da5dffd7bdc" Dec 27 08:28:57 crc kubenswrapper[4934]: E1227 08:28:57.156955 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77f5fd748da4d21e5b294de91a97165c9fd018cd107c77d73d219da5dffd7bdc\": container with ID starting with 77f5fd748da4d21e5b294de91a97165c9fd018cd107c77d73d219da5dffd7bdc not found: ID does not exist" containerID="77f5fd748da4d21e5b294de91a97165c9fd018cd107c77d73d219da5dffd7bdc" Dec 27 08:28:57 crc kubenswrapper[4934]: I1227 08:28:57.157007 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77f5fd748da4d21e5b294de91a97165c9fd018cd107c77d73d219da5dffd7bdc"} err="failed to get container status \"77f5fd748da4d21e5b294de91a97165c9fd018cd107c77d73d219da5dffd7bdc\": rpc error: code = NotFound desc = could not find container \"77f5fd748da4d21e5b294de91a97165c9fd018cd107c77d73d219da5dffd7bdc\": container with ID starting with 77f5fd748da4d21e5b294de91a97165c9fd018cd107c77d73d219da5dffd7bdc not found: ID does not exist" Dec 27 08:28:57 crc kubenswrapper[4934]: I1227 08:28:57.157044 4934 scope.go:117] "RemoveContainer" containerID="a13b4c2e3dcd9ca8871c29ca469ad825a2723eea2f4c6c841b13cc86dbae7611" Dec 27 08:28:57 crc kubenswrapper[4934]: E1227 08:28:57.157926 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a13b4c2e3dcd9ca8871c29ca469ad825a2723eea2f4c6c841b13cc86dbae7611\": container with ID starting with a13b4c2e3dcd9ca8871c29ca469ad825a2723eea2f4c6c841b13cc86dbae7611 not found: ID does not exist" containerID="a13b4c2e3dcd9ca8871c29ca469ad825a2723eea2f4c6c841b13cc86dbae7611" Dec 27 08:28:57 crc kubenswrapper[4934]: I1227 08:28:57.157961 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a13b4c2e3dcd9ca8871c29ca469ad825a2723eea2f4c6c841b13cc86dbae7611"} err="failed to get container status \"a13b4c2e3dcd9ca8871c29ca469ad825a2723eea2f4c6c841b13cc86dbae7611\": rpc error: code = NotFound desc = could not find container \"a13b4c2e3dcd9ca8871c29ca469ad825a2723eea2f4c6c841b13cc86dbae7611\": container with ID starting with a13b4c2e3dcd9ca8871c29ca469ad825a2723eea2f4c6c841b13cc86dbae7611 not found: ID does not exist" Dec 27 08:28:57 crc kubenswrapper[4934]: I1227 08:28:57.483731 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88a1181e-029e-4c3f-9b41-87f7500fb6b9" path="/var/lib/kubelet/pods/88a1181e-029e-4c3f-9b41-87f7500fb6b9/volumes" Dec 27 08:29:38 crc kubenswrapper[4934]: I1227 08:29:38.932464 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pbsw8"] Dec 27 08:29:38 crc kubenswrapper[4934]: E1227 08:29:38.933449 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88a1181e-029e-4c3f-9b41-87f7500fb6b9" containerName="extract-content" Dec 27 08:29:38 crc kubenswrapper[4934]: I1227 08:29:38.933463 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="88a1181e-029e-4c3f-9b41-87f7500fb6b9" containerName="extract-content" Dec 27 08:29:38 crc kubenswrapper[4934]: E1227 08:29:38.933497 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88a1181e-029e-4c3f-9b41-87f7500fb6b9" containerName="registry-server" Dec 27 08:29:38 crc kubenswrapper[4934]: I1227 08:29:38.933503 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="88a1181e-029e-4c3f-9b41-87f7500fb6b9" containerName="registry-server" Dec 27 08:29:38 crc kubenswrapper[4934]: E1227 08:29:38.933520 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88a1181e-029e-4c3f-9b41-87f7500fb6b9" containerName="extract-utilities" Dec 27 08:29:38 crc kubenswrapper[4934]: I1227 08:29:38.933526 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="88a1181e-029e-4c3f-9b41-87f7500fb6b9" containerName="extract-utilities" Dec 27 08:29:38 crc kubenswrapper[4934]: I1227 08:29:38.933770 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="88a1181e-029e-4c3f-9b41-87f7500fb6b9" containerName="registry-server" Dec 27 08:29:38 crc kubenswrapper[4934]: I1227 08:29:38.936136 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pbsw8" Dec 27 08:29:38 crc kubenswrapper[4934]: I1227 08:29:38.953220 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbsw8"] Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.075076 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c95cb32b-8a0b-4d09-a38e-25cf1b73b249-catalog-content\") pod \"redhat-marketplace-pbsw8\" (UID: \"c95cb32b-8a0b-4d09-a38e-25cf1b73b249\") " pod="openshift-marketplace/redhat-marketplace-pbsw8" Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.075605 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7css8\" (UniqueName: \"kubernetes.io/projected/c95cb32b-8a0b-4d09-a38e-25cf1b73b249-kube-api-access-7css8\") pod \"redhat-marketplace-pbsw8\" (UID: \"c95cb32b-8a0b-4d09-a38e-25cf1b73b249\") " pod="openshift-marketplace/redhat-marketplace-pbsw8" Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.075755 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c95cb32b-8a0b-4d09-a38e-25cf1b73b249-utilities\") pod \"redhat-marketplace-pbsw8\" (UID: \"c95cb32b-8a0b-4d09-a38e-25cf1b73b249\") " pod="openshift-marketplace/redhat-marketplace-pbsw8" Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.177969 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c95cb32b-8a0b-4d09-a38e-25cf1b73b249-catalog-content\") pod \"redhat-marketplace-pbsw8\" (UID: \"c95cb32b-8a0b-4d09-a38e-25cf1b73b249\") " pod="openshift-marketplace/redhat-marketplace-pbsw8" Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.178232 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7css8\" (UniqueName: \"kubernetes.io/projected/c95cb32b-8a0b-4d09-a38e-25cf1b73b249-kube-api-access-7css8\") pod \"redhat-marketplace-pbsw8\" (UID: \"c95cb32b-8a0b-4d09-a38e-25cf1b73b249\") " pod="openshift-marketplace/redhat-marketplace-pbsw8" Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.178283 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c95cb32b-8a0b-4d09-a38e-25cf1b73b249-utilities\") pod \"redhat-marketplace-pbsw8\" (UID: \"c95cb32b-8a0b-4d09-a38e-25cf1b73b249\") " pod="openshift-marketplace/redhat-marketplace-pbsw8" Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.178518 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c95cb32b-8a0b-4d09-a38e-25cf1b73b249-catalog-content\") pod \"redhat-marketplace-pbsw8\" (UID: \"c95cb32b-8a0b-4d09-a38e-25cf1b73b249\") " pod="openshift-marketplace/redhat-marketplace-pbsw8" Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.178681 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c95cb32b-8a0b-4d09-a38e-25cf1b73b249-utilities\") pod \"redhat-marketplace-pbsw8\" (UID: \"c95cb32b-8a0b-4d09-a38e-25cf1b73b249\") " pod="openshift-marketplace/redhat-marketplace-pbsw8" Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.197823 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7css8\" (UniqueName: \"kubernetes.io/projected/c95cb32b-8a0b-4d09-a38e-25cf1b73b249-kube-api-access-7css8\") pod \"redhat-marketplace-pbsw8\" (UID: \"c95cb32b-8a0b-4d09-a38e-25cf1b73b249\") " pod="openshift-marketplace/redhat-marketplace-pbsw8" Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.272681 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pbsw8" Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.308753 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-s4jfk"] Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.311891 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s4jfk" Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.325296 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s4jfk"] Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.485702 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7hdq\" (UniqueName: \"kubernetes.io/projected/5c02a72c-c148-47b6-bed7-fbca467ff35e-kube-api-access-b7hdq\") pod \"redhat-operators-s4jfk\" (UID: \"5c02a72c-c148-47b6-bed7-fbca467ff35e\") " pod="openshift-marketplace/redhat-operators-s4jfk" Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.486023 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c02a72c-c148-47b6-bed7-fbca467ff35e-utilities\") pod \"redhat-operators-s4jfk\" (UID: \"5c02a72c-c148-47b6-bed7-fbca467ff35e\") " pod="openshift-marketplace/redhat-operators-s4jfk" Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.486069 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c02a72c-c148-47b6-bed7-fbca467ff35e-catalog-content\") pod \"redhat-operators-s4jfk\" (UID: \"5c02a72c-c148-47b6-bed7-fbca467ff35e\") " pod="openshift-marketplace/redhat-operators-s4jfk" Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.588113 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c02a72c-c148-47b6-bed7-fbca467ff35e-catalog-content\") pod \"redhat-operators-s4jfk\" (UID: \"5c02a72c-c148-47b6-bed7-fbca467ff35e\") " pod="openshift-marketplace/redhat-operators-s4jfk" Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.588354 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7hdq\" (UniqueName: \"kubernetes.io/projected/5c02a72c-c148-47b6-bed7-fbca467ff35e-kube-api-access-b7hdq\") pod \"redhat-operators-s4jfk\" (UID: \"5c02a72c-c148-47b6-bed7-fbca467ff35e\") " pod="openshift-marketplace/redhat-operators-s4jfk" Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.588448 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c02a72c-c148-47b6-bed7-fbca467ff35e-utilities\") pod \"redhat-operators-s4jfk\" (UID: \"5c02a72c-c148-47b6-bed7-fbca467ff35e\") " pod="openshift-marketplace/redhat-operators-s4jfk" Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.588566 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c02a72c-c148-47b6-bed7-fbca467ff35e-catalog-content\") pod \"redhat-operators-s4jfk\" (UID: \"5c02a72c-c148-47b6-bed7-fbca467ff35e\") " pod="openshift-marketplace/redhat-operators-s4jfk" Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.588829 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c02a72c-c148-47b6-bed7-fbca467ff35e-utilities\") pod \"redhat-operators-s4jfk\" (UID: \"5c02a72c-c148-47b6-bed7-fbca467ff35e\") " pod="openshift-marketplace/redhat-operators-s4jfk" Dec 27 08:29:39 crc kubenswrapper[4934]: I1227 08:29:39.609718 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7hdq\" (UniqueName: \"kubernetes.io/projected/5c02a72c-c148-47b6-bed7-fbca467ff35e-kube-api-access-b7hdq\") pod \"redhat-operators-s4jfk\" (UID: \"5c02a72c-c148-47b6-bed7-fbca467ff35e\") " pod="openshift-marketplace/redhat-operators-s4jfk" Dec 27 08:29:40 crc kubenswrapper[4934]: I1227 08:29:40.982630 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s4jfk" Dec 27 08:29:41 crc kubenswrapper[4934]: I1227 08:29:41.137022 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbsw8"] Dec 27 08:29:41 crc kubenswrapper[4934]: I1227 08:29:41.675023 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s4jfk"] Dec 27 08:29:42 crc kubenswrapper[4934]: I1227 08:29:42.029022 4934 generic.go:334] "Generic (PLEG): container finished" podID="5c02a72c-c148-47b6-bed7-fbca467ff35e" containerID="e9996c58e6ebb7a4197e763659dd363c1ecb4c714478a2ddb88e720efa48dd2b" exitCode=0 Dec 27 08:29:42 crc kubenswrapper[4934]: I1227 08:29:42.029140 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4jfk" event={"ID":"5c02a72c-c148-47b6-bed7-fbca467ff35e","Type":"ContainerDied","Data":"e9996c58e6ebb7a4197e763659dd363c1ecb4c714478a2ddb88e720efa48dd2b"} Dec 27 08:29:42 crc kubenswrapper[4934]: I1227 08:29:42.031471 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4jfk" event={"ID":"5c02a72c-c148-47b6-bed7-fbca467ff35e","Type":"ContainerStarted","Data":"79ecce9b6a6e4fe8afd4b3a362ce0434434b88e157e9ce9c313a23b5343f0b43"} Dec 27 08:29:42 crc kubenswrapper[4934]: I1227 08:29:42.032814 4934 generic.go:334] "Generic (PLEG): container finished" podID="c95cb32b-8a0b-4d09-a38e-25cf1b73b249" containerID="a4585cb1a77284f5dd1179bb3d0eb19c7868343b278eea4ac8549de7c3ec0cc6" exitCode=0 Dec 27 08:29:42 crc kubenswrapper[4934]: I1227 08:29:42.032858 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbsw8" event={"ID":"c95cb32b-8a0b-4d09-a38e-25cf1b73b249","Type":"ContainerDied","Data":"a4585cb1a77284f5dd1179bb3d0eb19c7868343b278eea4ac8549de7c3ec0cc6"} Dec 27 08:29:42 crc kubenswrapper[4934]: I1227 08:29:42.032882 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbsw8" event={"ID":"c95cb32b-8a0b-4d09-a38e-25cf1b73b249","Type":"ContainerStarted","Data":"4f60e2a5aa8517276d40101c7423195c5f9dd87895516b73309a5941c22b25a7"} Dec 27 08:29:43 crc kubenswrapper[4934]: I1227 08:29:43.044917 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbsw8" event={"ID":"c95cb32b-8a0b-4d09-a38e-25cf1b73b249","Type":"ContainerStarted","Data":"600e39868422a431ca7c5f3822c592049ecc73e7309ff66232448460af617d51"} Dec 27 08:29:43 crc kubenswrapper[4934]: I1227 08:29:43.047967 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4jfk" event={"ID":"5c02a72c-c148-47b6-bed7-fbca467ff35e","Type":"ContainerStarted","Data":"c73931b92283b74a438c4603e457073d66184d13532f08946a87c77c6edd44b8"} Dec 27 08:29:43 crc kubenswrapper[4934]: E1227 08:29:43.805292 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc95cb32b_8a0b_4d09_a38e_25cf1b73b249.slice/crio-600e39868422a431ca7c5f3822c592049ecc73e7309ff66232448460af617d51.scope\": RecentStats: unable to find data in memory cache]" Dec 27 08:29:45 crc kubenswrapper[4934]: I1227 08:29:45.075946 4934 generic.go:334] "Generic (PLEG): container finished" podID="c95cb32b-8a0b-4d09-a38e-25cf1b73b249" containerID="600e39868422a431ca7c5f3822c592049ecc73e7309ff66232448460af617d51" exitCode=0 Dec 27 08:29:45 crc kubenswrapper[4934]: I1227 08:29:45.076031 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbsw8" event={"ID":"c95cb32b-8a0b-4d09-a38e-25cf1b73b249","Type":"ContainerDied","Data":"600e39868422a431ca7c5f3822c592049ecc73e7309ff66232448460af617d51"} Dec 27 08:29:47 crc kubenswrapper[4934]: I1227 08:29:47.101337 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbsw8" event={"ID":"c95cb32b-8a0b-4d09-a38e-25cf1b73b249","Type":"ContainerStarted","Data":"b0821e50ee6d684d12f7667be6bdfda9d84b5b10154843efca1f2fe3bf925d7d"} Dec 27 08:29:47 crc kubenswrapper[4934]: I1227 08:29:47.125251 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pbsw8" podStartSLOduration=5.422009846 podStartE2EDuration="9.125232498s" podCreationTimestamp="2025-12-27 08:29:38 +0000 UTC" firstStartedPulling="2025-12-27 08:29:42.034435685 +0000 UTC m=+2842.854876279" lastFinishedPulling="2025-12-27 08:29:45.737658297 +0000 UTC m=+2846.558098931" observedRunningTime="2025-12-27 08:29:47.116959033 +0000 UTC m=+2847.937399647" watchObservedRunningTime="2025-12-27 08:29:47.125232498 +0000 UTC m=+2847.945673092" Dec 27 08:29:48 crc kubenswrapper[4934]: I1227 08:29:48.119866 4934 generic.go:334] "Generic (PLEG): container finished" podID="5c02a72c-c148-47b6-bed7-fbca467ff35e" containerID="c73931b92283b74a438c4603e457073d66184d13532f08946a87c77c6edd44b8" exitCode=0 Dec 27 08:29:48 crc kubenswrapper[4934]: I1227 08:29:48.120248 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4jfk" event={"ID":"5c02a72c-c148-47b6-bed7-fbca467ff35e","Type":"ContainerDied","Data":"c73931b92283b74a438c4603e457073d66184d13532f08946a87c77c6edd44b8"} Dec 27 08:29:49 crc kubenswrapper[4934]: I1227 08:29:49.272953 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pbsw8" Dec 27 08:29:49 crc kubenswrapper[4934]: I1227 08:29:49.275141 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pbsw8" Dec 27 08:29:50 crc kubenswrapper[4934]: I1227 08:29:50.158012 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4jfk" event={"ID":"5c02a72c-c148-47b6-bed7-fbca467ff35e","Type":"ContainerStarted","Data":"7912857c59287fa286f4995f42f9ff15da80ee0eb69fc5cb50f7f9930ea6ebe1"} Dec 27 08:29:50 crc kubenswrapper[4934]: I1227 08:29:50.185331 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-s4jfk" podStartSLOduration=4.009938348 podStartE2EDuration="11.185309028s" podCreationTimestamp="2025-12-27 08:29:39 +0000 UTC" firstStartedPulling="2025-12-27 08:29:42.031122753 +0000 UTC m=+2842.851563387" lastFinishedPulling="2025-12-27 08:29:49.206493473 +0000 UTC m=+2850.026934067" observedRunningTime="2025-12-27 08:29:50.181538375 +0000 UTC m=+2851.001978999" watchObservedRunningTime="2025-12-27 08:29:50.185309028 +0000 UTC m=+2851.005749622" Dec 27 08:29:50 crc kubenswrapper[4934]: I1227 08:29:50.333665 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-pbsw8" podUID="c95cb32b-8a0b-4d09-a38e-25cf1b73b249" containerName="registry-server" probeResult="failure" output=< Dec 27 08:29:50 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 08:29:50 crc kubenswrapper[4934]: > Dec 27 08:29:50 crc kubenswrapper[4934]: I1227 08:29:50.984155 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-s4jfk" Dec 27 08:29:50 crc kubenswrapper[4934]: I1227 08:29:50.984415 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-s4jfk" Dec 27 08:29:52 crc kubenswrapper[4934]: I1227 08:29:52.051671 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-s4jfk" podUID="5c02a72c-c148-47b6-bed7-fbca467ff35e" containerName="registry-server" probeResult="failure" output=< Dec 27 08:29:52 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 08:29:52 crc kubenswrapper[4934]: > Dec 27 08:29:59 crc kubenswrapper[4934]: I1227 08:29:59.351141 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pbsw8" Dec 27 08:29:59 crc kubenswrapper[4934]: I1227 08:29:59.451602 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pbsw8" Dec 27 08:29:59 crc kubenswrapper[4934]: I1227 08:29:59.609933 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbsw8"] Dec 27 08:30:00 crc kubenswrapper[4934]: I1227 08:30:00.186663 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl"] Dec 27 08:30:00 crc kubenswrapper[4934]: I1227 08:30:00.189290 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl" Dec 27 08:30:00 crc kubenswrapper[4934]: I1227 08:30:00.192312 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 27 08:30:00 crc kubenswrapper[4934]: I1227 08:30:00.192841 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 27 08:30:00 crc kubenswrapper[4934]: I1227 08:30:00.200129 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl"] Dec 27 08:30:00 crc kubenswrapper[4934]: I1227 08:30:00.284390 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb7sh\" (UniqueName: \"kubernetes.io/projected/2f809b83-b833-46e0-9ff7-d52c567711e9-kube-api-access-fb7sh\") pod \"collect-profiles-29447070-hwzwl\" (UID: \"2f809b83-b833-46e0-9ff7-d52c567711e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl" Dec 27 08:30:00 crc kubenswrapper[4934]: I1227 08:30:00.284445 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f809b83-b833-46e0-9ff7-d52c567711e9-secret-volume\") pod \"collect-profiles-29447070-hwzwl\" (UID: \"2f809b83-b833-46e0-9ff7-d52c567711e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl" Dec 27 08:30:00 crc kubenswrapper[4934]: I1227 08:30:00.284777 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f809b83-b833-46e0-9ff7-d52c567711e9-config-volume\") pod \"collect-profiles-29447070-hwzwl\" (UID: \"2f809b83-b833-46e0-9ff7-d52c567711e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl" Dec 27 08:30:00 crc kubenswrapper[4934]: I1227 08:30:00.387450 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb7sh\" (UniqueName: \"kubernetes.io/projected/2f809b83-b833-46e0-9ff7-d52c567711e9-kube-api-access-fb7sh\") pod \"collect-profiles-29447070-hwzwl\" (UID: \"2f809b83-b833-46e0-9ff7-d52c567711e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl" Dec 27 08:30:00 crc kubenswrapper[4934]: I1227 08:30:00.387504 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f809b83-b833-46e0-9ff7-d52c567711e9-secret-volume\") pod \"collect-profiles-29447070-hwzwl\" (UID: \"2f809b83-b833-46e0-9ff7-d52c567711e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl" Dec 27 08:30:00 crc kubenswrapper[4934]: I1227 08:30:00.387666 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f809b83-b833-46e0-9ff7-d52c567711e9-config-volume\") pod \"collect-profiles-29447070-hwzwl\" (UID: \"2f809b83-b833-46e0-9ff7-d52c567711e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl" Dec 27 08:30:00 crc kubenswrapper[4934]: I1227 08:30:00.394132 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f809b83-b833-46e0-9ff7-d52c567711e9-config-volume\") pod \"collect-profiles-29447070-hwzwl\" (UID: \"2f809b83-b833-46e0-9ff7-d52c567711e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl" Dec 27 08:30:00 crc kubenswrapper[4934]: I1227 08:30:00.398058 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f809b83-b833-46e0-9ff7-d52c567711e9-secret-volume\") pod \"collect-profiles-29447070-hwzwl\" (UID: \"2f809b83-b833-46e0-9ff7-d52c567711e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl" Dec 27 08:30:00 crc kubenswrapper[4934]: I1227 08:30:00.406680 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb7sh\" (UniqueName: \"kubernetes.io/projected/2f809b83-b833-46e0-9ff7-d52c567711e9-kube-api-access-fb7sh\") pod \"collect-profiles-29447070-hwzwl\" (UID: \"2f809b83-b833-46e0-9ff7-d52c567711e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl" Dec 27 08:30:00 crc kubenswrapper[4934]: I1227 08:30:00.512591 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl" Dec 27 08:30:01 crc kubenswrapper[4934]: I1227 08:30:01.047924 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl"] Dec 27 08:30:01 crc kubenswrapper[4934]: I1227 08:30:01.084355 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-s4jfk" Dec 27 08:30:01 crc kubenswrapper[4934]: I1227 08:30:01.161440 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-s4jfk" Dec 27 08:30:01 crc kubenswrapper[4934]: I1227 08:30:01.307951 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl" event={"ID":"2f809b83-b833-46e0-9ff7-d52c567711e9","Type":"ContainerStarted","Data":"ace9f6c99252fcb22eb3a5770f78a4d70b8b8b2e3168809a11c3e4346e78edc6"} Dec 27 08:30:01 crc kubenswrapper[4934]: I1227 08:30:01.308030 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl" event={"ID":"2f809b83-b833-46e0-9ff7-d52c567711e9","Type":"ContainerStarted","Data":"108ec0807eb2244853c1653c96eead72fe326a730272a181fefceaa2bcd22fcc"} Dec 27 08:30:01 crc kubenswrapper[4934]: I1227 08:30:01.308504 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pbsw8" podUID="c95cb32b-8a0b-4d09-a38e-25cf1b73b249" containerName="registry-server" containerID="cri-o://b0821e50ee6d684d12f7667be6bdfda9d84b5b10154843efca1f2fe3bf925d7d" gracePeriod=2 Dec 27 08:30:01 crc kubenswrapper[4934]: I1227 08:30:01.335926 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl" podStartSLOduration=1.335906568 podStartE2EDuration="1.335906568s" podCreationTimestamp="2025-12-27 08:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 08:30:01.332987006 +0000 UTC m=+2862.153427610" watchObservedRunningTime="2025-12-27 08:30:01.335906568 +0000 UTC m=+2862.156347162" Dec 27 08:30:01 crc kubenswrapper[4934]: I1227 08:30:01.877214 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pbsw8" Dec 27 08:30:01 crc kubenswrapper[4934]: I1227 08:30:01.930351 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c95cb32b-8a0b-4d09-a38e-25cf1b73b249-catalog-content\") pod \"c95cb32b-8a0b-4d09-a38e-25cf1b73b249\" (UID: \"c95cb32b-8a0b-4d09-a38e-25cf1b73b249\") " Dec 27 08:30:01 crc kubenswrapper[4934]: I1227 08:30:01.930826 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7css8\" (UniqueName: \"kubernetes.io/projected/c95cb32b-8a0b-4d09-a38e-25cf1b73b249-kube-api-access-7css8\") pod \"c95cb32b-8a0b-4d09-a38e-25cf1b73b249\" (UID: \"c95cb32b-8a0b-4d09-a38e-25cf1b73b249\") " Dec 27 08:30:01 crc kubenswrapper[4934]: I1227 08:30:01.930882 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c95cb32b-8a0b-4d09-a38e-25cf1b73b249-utilities\") pod \"c95cb32b-8a0b-4d09-a38e-25cf1b73b249\" (UID: \"c95cb32b-8a0b-4d09-a38e-25cf1b73b249\") " Dec 27 08:30:01 crc kubenswrapper[4934]: I1227 08:30:01.931595 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c95cb32b-8a0b-4d09-a38e-25cf1b73b249-utilities" (OuterVolumeSpecName: "utilities") pod "c95cb32b-8a0b-4d09-a38e-25cf1b73b249" (UID: "c95cb32b-8a0b-4d09-a38e-25cf1b73b249"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:30:01 crc kubenswrapper[4934]: I1227 08:30:01.939292 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c95cb32b-8a0b-4d09-a38e-25cf1b73b249-kube-api-access-7css8" (OuterVolumeSpecName: "kube-api-access-7css8") pod "c95cb32b-8a0b-4d09-a38e-25cf1b73b249" (UID: "c95cb32b-8a0b-4d09-a38e-25cf1b73b249"). InnerVolumeSpecName "kube-api-access-7css8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:30:01 crc kubenswrapper[4934]: I1227 08:30:01.957508 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c95cb32b-8a0b-4d09-a38e-25cf1b73b249-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c95cb32b-8a0b-4d09-a38e-25cf1b73b249" (UID: "c95cb32b-8a0b-4d09-a38e-25cf1b73b249"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.016688 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s4jfk"] Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.033668 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c95cb32b-8a0b-4d09-a38e-25cf1b73b249-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.033714 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7css8\" (UniqueName: \"kubernetes.io/projected/c95cb32b-8a0b-4d09-a38e-25cf1b73b249-kube-api-access-7css8\") on node \"crc\" DevicePath \"\"" Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.033732 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c95cb32b-8a0b-4d09-a38e-25cf1b73b249-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.324524 4934 generic.go:334] "Generic (PLEG): container finished" podID="c95cb32b-8a0b-4d09-a38e-25cf1b73b249" containerID="b0821e50ee6d684d12f7667be6bdfda9d84b5b10154843efca1f2fe3bf925d7d" exitCode=0 Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.324591 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbsw8" event={"ID":"c95cb32b-8a0b-4d09-a38e-25cf1b73b249","Type":"ContainerDied","Data":"b0821e50ee6d684d12f7667be6bdfda9d84b5b10154843efca1f2fe3bf925d7d"} Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.324624 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pbsw8" Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.324722 4934 scope.go:117] "RemoveContainer" containerID="b0821e50ee6d684d12f7667be6bdfda9d84b5b10154843efca1f2fe3bf925d7d" Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.324696 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pbsw8" event={"ID":"c95cb32b-8a0b-4d09-a38e-25cf1b73b249","Type":"ContainerDied","Data":"4f60e2a5aa8517276d40101c7423195c5f9dd87895516b73309a5941c22b25a7"} Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.327453 4934 generic.go:334] "Generic (PLEG): container finished" podID="2f809b83-b833-46e0-9ff7-d52c567711e9" containerID="ace9f6c99252fcb22eb3a5770f78a4d70b8b8b2e3168809a11c3e4346e78edc6" exitCode=0 Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.327678 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-s4jfk" podUID="5c02a72c-c148-47b6-bed7-fbca467ff35e" containerName="registry-server" containerID="cri-o://7912857c59287fa286f4995f42f9ff15da80ee0eb69fc5cb50f7f9930ea6ebe1" gracePeriod=2 Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.327708 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl" event={"ID":"2f809b83-b833-46e0-9ff7-d52c567711e9","Type":"ContainerDied","Data":"ace9f6c99252fcb22eb3a5770f78a4d70b8b8b2e3168809a11c3e4346e78edc6"} Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.368409 4934 scope.go:117] "RemoveContainer" containerID="600e39868422a431ca7c5f3822c592049ecc73e7309ff66232448460af617d51" Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.392352 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbsw8"] Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.408123 4934 scope.go:117] "RemoveContainer" containerID="a4585cb1a77284f5dd1179bb3d0eb19c7868343b278eea4ac8549de7c3ec0cc6" Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.410996 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pbsw8"] Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.616871 4934 scope.go:117] "RemoveContainer" containerID="b0821e50ee6d684d12f7667be6bdfda9d84b5b10154843efca1f2fe3bf925d7d" Dec 27 08:30:02 crc kubenswrapper[4934]: E1227 08:30:02.618962 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0821e50ee6d684d12f7667be6bdfda9d84b5b10154843efca1f2fe3bf925d7d\": container with ID starting with b0821e50ee6d684d12f7667be6bdfda9d84b5b10154843efca1f2fe3bf925d7d not found: ID does not exist" containerID="b0821e50ee6d684d12f7667be6bdfda9d84b5b10154843efca1f2fe3bf925d7d" Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.619041 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0821e50ee6d684d12f7667be6bdfda9d84b5b10154843efca1f2fe3bf925d7d"} err="failed to get container status \"b0821e50ee6d684d12f7667be6bdfda9d84b5b10154843efca1f2fe3bf925d7d\": rpc error: code = NotFound desc = could not find container \"b0821e50ee6d684d12f7667be6bdfda9d84b5b10154843efca1f2fe3bf925d7d\": container with ID starting with b0821e50ee6d684d12f7667be6bdfda9d84b5b10154843efca1f2fe3bf925d7d not found: ID does not exist" Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.619096 4934 scope.go:117] "RemoveContainer" containerID="600e39868422a431ca7c5f3822c592049ecc73e7309ff66232448460af617d51" Dec 27 08:30:02 crc kubenswrapper[4934]: E1227 08:30:02.620713 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"600e39868422a431ca7c5f3822c592049ecc73e7309ff66232448460af617d51\": container with ID starting with 600e39868422a431ca7c5f3822c592049ecc73e7309ff66232448460af617d51 not found: ID does not exist" containerID="600e39868422a431ca7c5f3822c592049ecc73e7309ff66232448460af617d51" Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.620764 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"600e39868422a431ca7c5f3822c592049ecc73e7309ff66232448460af617d51"} err="failed to get container status \"600e39868422a431ca7c5f3822c592049ecc73e7309ff66232448460af617d51\": rpc error: code = NotFound desc = could not find container \"600e39868422a431ca7c5f3822c592049ecc73e7309ff66232448460af617d51\": container with ID starting with 600e39868422a431ca7c5f3822c592049ecc73e7309ff66232448460af617d51 not found: ID does not exist" Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.620797 4934 scope.go:117] "RemoveContainer" containerID="a4585cb1a77284f5dd1179bb3d0eb19c7868343b278eea4ac8549de7c3ec0cc6" Dec 27 08:30:02 crc kubenswrapper[4934]: E1227 08:30:02.621222 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4585cb1a77284f5dd1179bb3d0eb19c7868343b278eea4ac8549de7c3ec0cc6\": container with ID starting with a4585cb1a77284f5dd1179bb3d0eb19c7868343b278eea4ac8549de7c3ec0cc6 not found: ID does not exist" containerID="a4585cb1a77284f5dd1179bb3d0eb19c7868343b278eea4ac8549de7c3ec0cc6" Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.621251 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4585cb1a77284f5dd1179bb3d0eb19c7868343b278eea4ac8549de7c3ec0cc6"} err="failed to get container status \"a4585cb1a77284f5dd1179bb3d0eb19c7868343b278eea4ac8549de7c3ec0cc6\": rpc error: code = NotFound desc = could not find container \"a4585cb1a77284f5dd1179bb3d0eb19c7868343b278eea4ac8549de7c3ec0cc6\": container with ID starting with a4585cb1a77284f5dd1179bb3d0eb19c7868343b278eea4ac8549de7c3ec0cc6 not found: ID does not exist" Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.922428 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s4jfk" Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.965552 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7hdq\" (UniqueName: \"kubernetes.io/projected/5c02a72c-c148-47b6-bed7-fbca467ff35e-kube-api-access-b7hdq\") pod \"5c02a72c-c148-47b6-bed7-fbca467ff35e\" (UID: \"5c02a72c-c148-47b6-bed7-fbca467ff35e\") " Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.965623 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c02a72c-c148-47b6-bed7-fbca467ff35e-catalog-content\") pod \"5c02a72c-c148-47b6-bed7-fbca467ff35e\" (UID: \"5c02a72c-c148-47b6-bed7-fbca467ff35e\") " Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.965773 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c02a72c-c148-47b6-bed7-fbca467ff35e-utilities\") pod \"5c02a72c-c148-47b6-bed7-fbca467ff35e\" (UID: \"5c02a72c-c148-47b6-bed7-fbca467ff35e\") " Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.967115 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c02a72c-c148-47b6-bed7-fbca467ff35e-utilities" (OuterVolumeSpecName: "utilities") pod "5c02a72c-c148-47b6-bed7-fbca467ff35e" (UID: "5c02a72c-c148-47b6-bed7-fbca467ff35e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:30:02 crc kubenswrapper[4934]: I1227 08:30:02.979058 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c02a72c-c148-47b6-bed7-fbca467ff35e-kube-api-access-b7hdq" (OuterVolumeSpecName: "kube-api-access-b7hdq") pod "5c02a72c-c148-47b6-bed7-fbca467ff35e" (UID: "5c02a72c-c148-47b6-bed7-fbca467ff35e"). InnerVolumeSpecName "kube-api-access-b7hdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.068771 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c02a72c-c148-47b6-bed7-fbca467ff35e-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.068835 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7hdq\" (UniqueName: \"kubernetes.io/projected/5c02a72c-c148-47b6-bed7-fbca467ff35e-kube-api-access-b7hdq\") on node \"crc\" DevicePath \"\"" Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.102632 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c02a72c-c148-47b6-bed7-fbca467ff35e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5c02a72c-c148-47b6-bed7-fbca467ff35e" (UID: "5c02a72c-c148-47b6-bed7-fbca467ff35e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.171804 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c02a72c-c148-47b6-bed7-fbca467ff35e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.360477 4934 generic.go:334] "Generic (PLEG): container finished" podID="5c02a72c-c148-47b6-bed7-fbca467ff35e" containerID="7912857c59287fa286f4995f42f9ff15da80ee0eb69fc5cb50f7f9930ea6ebe1" exitCode=0 Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.360587 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4jfk" event={"ID":"5c02a72c-c148-47b6-bed7-fbca467ff35e","Type":"ContainerDied","Data":"7912857c59287fa286f4995f42f9ff15da80ee0eb69fc5cb50f7f9930ea6ebe1"} Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.360664 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s4jfk" event={"ID":"5c02a72c-c148-47b6-bed7-fbca467ff35e","Type":"ContainerDied","Data":"79ecce9b6a6e4fe8afd4b3a362ce0434434b88e157e9ce9c313a23b5343f0b43"} Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.360755 4934 scope.go:117] "RemoveContainer" containerID="7912857c59287fa286f4995f42f9ff15da80ee0eb69fc5cb50f7f9930ea6ebe1" Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.362203 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s4jfk" Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.421169 4934 scope.go:117] "RemoveContainer" containerID="c73931b92283b74a438c4603e457073d66184d13532f08946a87c77c6edd44b8" Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.442966 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s4jfk"] Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.459944 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-s4jfk"] Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.516399 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c02a72c-c148-47b6-bed7-fbca467ff35e" path="/var/lib/kubelet/pods/5c02a72c-c148-47b6-bed7-fbca467ff35e/volumes" Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.517681 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c95cb32b-8a0b-4d09-a38e-25cf1b73b249" path="/var/lib/kubelet/pods/c95cb32b-8a0b-4d09-a38e-25cf1b73b249/volumes" Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.537952 4934 scope.go:117] "RemoveContainer" containerID="e9996c58e6ebb7a4197e763659dd363c1ecb4c714478a2ddb88e720efa48dd2b" Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.566439 4934 scope.go:117] "RemoveContainer" containerID="7912857c59287fa286f4995f42f9ff15da80ee0eb69fc5cb50f7f9930ea6ebe1" Dec 27 08:30:03 crc kubenswrapper[4934]: E1227 08:30:03.566957 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7912857c59287fa286f4995f42f9ff15da80ee0eb69fc5cb50f7f9930ea6ebe1\": container with ID starting with 7912857c59287fa286f4995f42f9ff15da80ee0eb69fc5cb50f7f9930ea6ebe1 not found: ID does not exist" containerID="7912857c59287fa286f4995f42f9ff15da80ee0eb69fc5cb50f7f9930ea6ebe1" Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.567001 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7912857c59287fa286f4995f42f9ff15da80ee0eb69fc5cb50f7f9930ea6ebe1"} err="failed to get container status \"7912857c59287fa286f4995f42f9ff15da80ee0eb69fc5cb50f7f9930ea6ebe1\": rpc error: code = NotFound desc = could not find container \"7912857c59287fa286f4995f42f9ff15da80ee0eb69fc5cb50f7f9930ea6ebe1\": container with ID starting with 7912857c59287fa286f4995f42f9ff15da80ee0eb69fc5cb50f7f9930ea6ebe1 not found: ID does not exist" Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.567030 4934 scope.go:117] "RemoveContainer" containerID="c73931b92283b74a438c4603e457073d66184d13532f08946a87c77c6edd44b8" Dec 27 08:30:03 crc kubenswrapper[4934]: E1227 08:30:03.567515 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c73931b92283b74a438c4603e457073d66184d13532f08946a87c77c6edd44b8\": container with ID starting with c73931b92283b74a438c4603e457073d66184d13532f08946a87c77c6edd44b8 not found: ID does not exist" containerID="c73931b92283b74a438c4603e457073d66184d13532f08946a87c77c6edd44b8" Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.567542 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c73931b92283b74a438c4603e457073d66184d13532f08946a87c77c6edd44b8"} err="failed to get container status \"c73931b92283b74a438c4603e457073d66184d13532f08946a87c77c6edd44b8\": rpc error: code = NotFound desc = could not find container \"c73931b92283b74a438c4603e457073d66184d13532f08946a87c77c6edd44b8\": container with ID starting with c73931b92283b74a438c4603e457073d66184d13532f08946a87c77c6edd44b8 not found: ID does not exist" Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.567556 4934 scope.go:117] "RemoveContainer" containerID="e9996c58e6ebb7a4197e763659dd363c1ecb4c714478a2ddb88e720efa48dd2b" Dec 27 08:30:03 crc kubenswrapper[4934]: E1227 08:30:03.567836 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9996c58e6ebb7a4197e763659dd363c1ecb4c714478a2ddb88e720efa48dd2b\": container with ID starting with e9996c58e6ebb7a4197e763659dd363c1ecb4c714478a2ddb88e720efa48dd2b not found: ID does not exist" containerID="e9996c58e6ebb7a4197e763659dd363c1ecb4c714478a2ddb88e720efa48dd2b" Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.567854 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9996c58e6ebb7a4197e763659dd363c1ecb4c714478a2ddb88e720efa48dd2b"} err="failed to get container status \"e9996c58e6ebb7a4197e763659dd363c1ecb4c714478a2ddb88e720efa48dd2b\": rpc error: code = NotFound desc = could not find container \"e9996c58e6ebb7a4197e763659dd363c1ecb4c714478a2ddb88e720efa48dd2b\": container with ID starting with e9996c58e6ebb7a4197e763659dd363c1ecb4c714478a2ddb88e720efa48dd2b not found: ID does not exist" Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.805479 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl" Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.897411 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f809b83-b833-46e0-9ff7-d52c567711e9-config-volume\") pod \"2f809b83-b833-46e0-9ff7-d52c567711e9\" (UID: \"2f809b83-b833-46e0-9ff7-d52c567711e9\") " Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.897874 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fb7sh\" (UniqueName: \"kubernetes.io/projected/2f809b83-b833-46e0-9ff7-d52c567711e9-kube-api-access-fb7sh\") pod \"2f809b83-b833-46e0-9ff7-d52c567711e9\" (UID: \"2f809b83-b833-46e0-9ff7-d52c567711e9\") " Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.898010 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f809b83-b833-46e0-9ff7-d52c567711e9-secret-volume\") pod \"2f809b83-b833-46e0-9ff7-d52c567711e9\" (UID: \"2f809b83-b833-46e0-9ff7-d52c567711e9\") " Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.898155 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f809b83-b833-46e0-9ff7-d52c567711e9-config-volume" (OuterVolumeSpecName: "config-volume") pod "2f809b83-b833-46e0-9ff7-d52c567711e9" (UID: "2f809b83-b833-46e0-9ff7-d52c567711e9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.899072 4934 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2f809b83-b833-46e0-9ff7-d52c567711e9-config-volume\") on node \"crc\" DevicePath \"\"" Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.902013 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f809b83-b833-46e0-9ff7-d52c567711e9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2f809b83-b833-46e0-9ff7-d52c567711e9" (UID: "2f809b83-b833-46e0-9ff7-d52c567711e9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:30:03 crc kubenswrapper[4934]: I1227 08:30:03.902277 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f809b83-b833-46e0-9ff7-d52c567711e9-kube-api-access-fb7sh" (OuterVolumeSpecName: "kube-api-access-fb7sh") pod "2f809b83-b833-46e0-9ff7-d52c567711e9" (UID: "2f809b83-b833-46e0-9ff7-d52c567711e9"). InnerVolumeSpecName "kube-api-access-fb7sh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:30:04 crc kubenswrapper[4934]: I1227 08:30:04.001656 4934 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2f809b83-b833-46e0-9ff7-d52c567711e9-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 27 08:30:04 crc kubenswrapper[4934]: I1227 08:30:04.001692 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fb7sh\" (UniqueName: \"kubernetes.io/projected/2f809b83-b833-46e0-9ff7-d52c567711e9-kube-api-access-fb7sh\") on node \"crc\" DevicePath \"\"" Dec 27 08:30:04 crc kubenswrapper[4934]: I1227 08:30:04.384630 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl" event={"ID":"2f809b83-b833-46e0-9ff7-d52c567711e9","Type":"ContainerDied","Data":"108ec0807eb2244853c1653c96eead72fe326a730272a181fefceaa2bcd22fcc"} Dec 27 08:30:04 crc kubenswrapper[4934]: I1227 08:30:04.384690 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="108ec0807eb2244853c1653c96eead72fe326a730272a181fefceaa2bcd22fcc" Dec 27 08:30:04 crc kubenswrapper[4934]: I1227 08:30:04.384786 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl" Dec 27 08:30:04 crc kubenswrapper[4934]: I1227 08:30:04.431878 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs"] Dec 27 08:30:04 crc kubenswrapper[4934]: I1227 08:30:04.452964 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447025-g88fs"] Dec 27 08:30:05 crc kubenswrapper[4934]: I1227 08:30:05.484367 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20da3be4-5db1-43ec-9b86-5ffae5145bf8" path="/var/lib/kubelet/pods/20da3be4-5db1-43ec-9b86-5ffae5145bf8/volumes" Dec 27 08:30:28 crc kubenswrapper[4934]: I1227 08:30:28.656701 4934 scope.go:117] "RemoveContainer" containerID="6852ddecc1458ae49f48e96d8d6f3450fe964be22e0c1a48aa7b5a131b953f19" Dec 27 08:30:45 crc kubenswrapper[4934]: I1227 08:30:45.330071 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:30:45 crc kubenswrapper[4934]: I1227 08:30:45.330572 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:31:15 crc kubenswrapper[4934]: I1227 08:31:15.330139 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:31:15 crc kubenswrapper[4934]: I1227 08:31:15.330771 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:31:44 crc kubenswrapper[4934]: I1227 08:31:44.917809 4934 generic.go:334] "Generic (PLEG): container finished" podID="2f4288f0-95f7-4b96-8066-e8b892d401ae" containerID="f9dc6bcada6d7de5b3b06d56d9ddec2526649242476946d4d992e13681c530d6" exitCode=0 Dec 27 08:31:44 crc kubenswrapper[4934]: I1227 08:31:44.917937 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" event={"ID":"2f4288f0-95f7-4b96-8066-e8b892d401ae","Type":"ContainerDied","Data":"f9dc6bcada6d7de5b3b06d56d9ddec2526649242476946d4d992e13681c530d6"} Dec 27 08:31:45 crc kubenswrapper[4934]: I1227 08:31:45.330189 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:31:45 crc kubenswrapper[4934]: I1227 08:31:45.330529 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:31:45 crc kubenswrapper[4934]: I1227 08:31:45.330717 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 08:31:45 crc kubenswrapper[4934]: I1227 08:31:45.332051 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 08:31:45 crc kubenswrapper[4934]: I1227 08:31:45.332287 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" gracePeriod=600 Dec 27 08:31:45 crc kubenswrapper[4934]: I1227 08:31:45.932177 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" exitCode=0 Dec 27 08:31:45 crc kubenswrapper[4934]: I1227 08:31:45.932212 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b"} Dec 27 08:31:45 crc kubenswrapper[4934]: I1227 08:31:45.932624 4934 scope.go:117] "RemoveContainer" containerID="0b5192c2223a1f8a1927e3908026b1994ec18d42dcd71c448b0446cd8d833be2" Dec 27 08:31:46 crc kubenswrapper[4934]: E1227 08:31:46.055943 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.543586 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.621323 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-inventory\") pod \"2f4288f0-95f7-4b96-8066-e8b892d401ae\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.621683 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-migration-ssh-key-0\") pod \"2f4288f0-95f7-4b96-8066-e8b892d401ae\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.621836 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-combined-ca-bundle\") pod \"2f4288f0-95f7-4b96-8066-e8b892d401ae\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.622007 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-cell1-compute-config-0\") pod \"2f4288f0-95f7-4b96-8066-e8b892d401ae\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.622134 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-cell1-compute-config-1\") pod \"2f4288f0-95f7-4b96-8066-e8b892d401ae\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.622312 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-ssh-key\") pod \"2f4288f0-95f7-4b96-8066-e8b892d401ae\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.622500 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4v79\" (UniqueName: \"kubernetes.io/projected/2f4288f0-95f7-4b96-8066-e8b892d401ae-kube-api-access-w4v79\") pod \"2f4288f0-95f7-4b96-8066-e8b892d401ae\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.622592 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-migration-ssh-key-1\") pod \"2f4288f0-95f7-4b96-8066-e8b892d401ae\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.622713 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-extra-config-0\") pod \"2f4288f0-95f7-4b96-8066-e8b892d401ae\" (UID: \"2f4288f0-95f7-4b96-8066-e8b892d401ae\") " Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.633241 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f4288f0-95f7-4b96-8066-e8b892d401ae-kube-api-access-w4v79" (OuterVolumeSpecName: "kube-api-access-w4v79") pod "2f4288f0-95f7-4b96-8066-e8b892d401ae" (UID: "2f4288f0-95f7-4b96-8066-e8b892d401ae"). InnerVolumeSpecName "kube-api-access-w4v79". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.638235 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "2f4288f0-95f7-4b96-8066-e8b892d401ae" (UID: "2f4288f0-95f7-4b96-8066-e8b892d401ae"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.674262 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "2f4288f0-95f7-4b96-8066-e8b892d401ae" (UID: "2f4288f0-95f7-4b96-8066-e8b892d401ae"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.675629 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "2f4288f0-95f7-4b96-8066-e8b892d401ae" (UID: "2f4288f0-95f7-4b96-8066-e8b892d401ae"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.677332 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "2f4288f0-95f7-4b96-8066-e8b892d401ae" (UID: "2f4288f0-95f7-4b96-8066-e8b892d401ae"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.679616 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "2f4288f0-95f7-4b96-8066-e8b892d401ae" (UID: "2f4288f0-95f7-4b96-8066-e8b892d401ae"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.697376 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-inventory" (OuterVolumeSpecName: "inventory") pod "2f4288f0-95f7-4b96-8066-e8b892d401ae" (UID: "2f4288f0-95f7-4b96-8066-e8b892d401ae"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.709031 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2f4288f0-95f7-4b96-8066-e8b892d401ae" (UID: "2f4288f0-95f7-4b96-8066-e8b892d401ae"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.709300 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "2f4288f0-95f7-4b96-8066-e8b892d401ae" (UID: "2f4288f0-95f7-4b96-8066-e8b892d401ae"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.726984 4934 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.727040 4934 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.727059 4934 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.727077 4934 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.727190 4934 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.727208 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4v79\" (UniqueName: \"kubernetes.io/projected/2f4288f0-95f7-4b96-8066-e8b892d401ae-kube-api-access-w4v79\") on node \"crc\" DevicePath \"\"" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.727225 4934 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.727242 4934 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/2f4288f0-95f7-4b96-8066-e8b892d401ae-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.727261 4934 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f4288f0-95f7-4b96-8066-e8b892d401ae-inventory\") on node \"crc\" DevicePath \"\"" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.951425 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:31:46 crc kubenswrapper[4934]: E1227 08:31:46.952127 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.958941 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" event={"ID":"2f4288f0-95f7-4b96-8066-e8b892d401ae","Type":"ContainerDied","Data":"ebd82b0d988be348057fe3ff273529d3991ade7822c9010d1f2e68a1bbf6c900"} Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.959001 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebd82b0d988be348057fe3ff273529d3991ade7822c9010d1f2e68a1bbf6c900" Dec 27 08:31:46 crc kubenswrapper[4934]: I1227 08:31:46.959018 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-qb7m2" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.101017 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm"] Dec 27 08:31:47 crc kubenswrapper[4934]: E1227 08:31:47.102146 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f809b83-b833-46e0-9ff7-d52c567711e9" containerName="collect-profiles" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.102164 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f809b83-b833-46e0-9ff7-d52c567711e9" containerName="collect-profiles" Dec 27 08:31:47 crc kubenswrapper[4934]: E1227 08:31:47.102183 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c02a72c-c148-47b6-bed7-fbca467ff35e" containerName="extract-utilities" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.102189 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c02a72c-c148-47b6-bed7-fbca467ff35e" containerName="extract-utilities" Dec 27 08:31:47 crc kubenswrapper[4934]: E1227 08:31:47.102214 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c02a72c-c148-47b6-bed7-fbca467ff35e" containerName="extract-content" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.102220 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c02a72c-c148-47b6-bed7-fbca467ff35e" containerName="extract-content" Dec 27 08:31:47 crc kubenswrapper[4934]: E1227 08:31:47.102245 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f4288f0-95f7-4b96-8066-e8b892d401ae" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.102251 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f4288f0-95f7-4b96-8066-e8b892d401ae" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 27 08:31:47 crc kubenswrapper[4934]: E1227 08:31:47.102266 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c95cb32b-8a0b-4d09-a38e-25cf1b73b249" containerName="registry-server" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.102272 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c95cb32b-8a0b-4d09-a38e-25cf1b73b249" containerName="registry-server" Dec 27 08:31:47 crc kubenswrapper[4934]: E1227 08:31:47.102282 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c02a72c-c148-47b6-bed7-fbca467ff35e" containerName="registry-server" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.102289 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c02a72c-c148-47b6-bed7-fbca467ff35e" containerName="registry-server" Dec 27 08:31:47 crc kubenswrapper[4934]: E1227 08:31:47.102303 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c95cb32b-8a0b-4d09-a38e-25cf1b73b249" containerName="extract-content" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.102308 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c95cb32b-8a0b-4d09-a38e-25cf1b73b249" containerName="extract-content" Dec 27 08:31:47 crc kubenswrapper[4934]: E1227 08:31:47.102320 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c95cb32b-8a0b-4d09-a38e-25cf1b73b249" containerName="extract-utilities" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.102325 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c95cb32b-8a0b-4d09-a38e-25cf1b73b249" containerName="extract-utilities" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.102533 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c95cb32b-8a0b-4d09-a38e-25cf1b73b249" containerName="registry-server" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.102549 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f4288f0-95f7-4b96-8066-e8b892d401ae" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.102565 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f809b83-b833-46e0-9ff7-d52c567711e9" containerName="collect-profiles" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.102591 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c02a72c-c148-47b6-bed7-fbca467ff35e" containerName="registry-server" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.103951 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.106386 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.106455 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.106515 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.106528 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-98ddn" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.106534 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.115095 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm"] Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.240115 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.240200 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.240244 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hczqx\" (UniqueName: \"kubernetes.io/projected/b87a1e52-c2d2-4384-97ee-c3cd68e90386-kube-api-access-hczqx\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.240318 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.240400 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.240751 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.240810 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.343009 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.343275 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.343380 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hczqx\" (UniqueName: \"kubernetes.io/projected/b87a1e52-c2d2-4384-97ee-c3cd68e90386-kube-api-access-hczqx\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.343766 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.343893 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.344062 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.344209 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.348954 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.349111 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.349960 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.350416 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.364563 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.366345 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hczqx\" (UniqueName: \"kubernetes.io/projected/b87a1e52-c2d2-4384-97ee-c3cd68e90386-kube-api-access-hczqx\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.378209 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:47 crc kubenswrapper[4934]: I1227 08:31:47.427012 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:31:48 crc kubenswrapper[4934]: I1227 08:31:48.048446 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm"] Dec 27 08:31:48 crc kubenswrapper[4934]: I1227 08:31:48.988801 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" event={"ID":"b87a1e52-c2d2-4384-97ee-c3cd68e90386","Type":"ContainerStarted","Data":"c0c1ad6c580ad447cd2b062b7e8aa03701b0252c65fbb50cf72d424db89ac31f"} Dec 27 08:31:51 crc kubenswrapper[4934]: I1227 08:31:51.022501 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" event={"ID":"b87a1e52-c2d2-4384-97ee-c3cd68e90386","Type":"ContainerStarted","Data":"1e16dce6eeccd386f609700a03cae5f999c01d7acf06f512d573d2c57940d8a0"} Dec 27 08:32:01 crc kubenswrapper[4934]: I1227 08:32:01.468610 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:32:01 crc kubenswrapper[4934]: E1227 08:32:01.470306 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:32:16 crc kubenswrapper[4934]: I1227 08:32:16.467819 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:32:16 crc kubenswrapper[4934]: E1227 08:32:16.468754 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:32:31 crc kubenswrapper[4934]: I1227 08:32:31.468278 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:32:31 crc kubenswrapper[4934]: E1227 08:32:31.469624 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:32:42 crc kubenswrapper[4934]: I1227 08:32:42.468173 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:32:42 crc kubenswrapper[4934]: E1227 08:32:42.469246 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:32:55 crc kubenswrapper[4934]: I1227 08:32:55.468212 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:32:55 crc kubenswrapper[4934]: E1227 08:32:55.469963 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:33:09 crc kubenswrapper[4934]: I1227 08:33:09.468032 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:33:09 crc kubenswrapper[4934]: E1227 08:33:09.468916 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:33:23 crc kubenswrapper[4934]: I1227 08:33:23.467761 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:33:23 crc kubenswrapper[4934]: E1227 08:33:23.468730 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:33:35 crc kubenswrapper[4934]: I1227 08:33:35.468301 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:33:35 crc kubenswrapper[4934]: E1227 08:33:35.469537 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:33:47 crc kubenswrapper[4934]: I1227 08:33:47.470655 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:33:47 crc kubenswrapper[4934]: E1227 08:33:47.471942 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:34:01 crc kubenswrapper[4934]: I1227 08:34:01.468823 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:34:01 crc kubenswrapper[4934]: E1227 08:34:01.470143 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:34:16 crc kubenswrapper[4934]: I1227 08:34:16.468108 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:34:16 crc kubenswrapper[4934]: E1227 08:34:16.468767 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:34:29 crc kubenswrapper[4934]: I1227 08:34:29.475230 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:34:29 crc kubenswrapper[4934]: E1227 08:34:29.476561 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:34:40 crc kubenswrapper[4934]: I1227 08:34:40.467729 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:34:40 crc kubenswrapper[4934]: E1227 08:34:40.468633 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:34:43 crc kubenswrapper[4934]: I1227 08:34:43.185125 4934 generic.go:334] "Generic (PLEG): container finished" podID="b87a1e52-c2d2-4384-97ee-c3cd68e90386" containerID="1e16dce6eeccd386f609700a03cae5f999c01d7acf06f512d573d2c57940d8a0" exitCode=0 Dec 27 08:34:43 crc kubenswrapper[4934]: I1227 08:34:43.185208 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" event={"ID":"b87a1e52-c2d2-4384-97ee-c3cd68e90386","Type":"ContainerDied","Data":"1e16dce6eeccd386f609700a03cae5f999c01d7acf06f512d573d2c57940d8a0"} Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.816173 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.870921 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ssh-key\") pod \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.871019 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-telemetry-combined-ca-bundle\") pod \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.871147 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ceilometer-compute-config-data-2\") pod \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.871178 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ceilometer-compute-config-data-0\") pod \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.871199 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-inventory\") pod \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.871307 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hczqx\" (UniqueName: \"kubernetes.io/projected/b87a1e52-c2d2-4384-97ee-c3cd68e90386-kube-api-access-hczqx\") pod \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.871350 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ceilometer-compute-config-data-1\") pod \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\" (UID: \"b87a1e52-c2d2-4384-97ee-c3cd68e90386\") " Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.886424 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b87a1e52-c2d2-4384-97ee-c3cd68e90386-kube-api-access-hczqx" (OuterVolumeSpecName: "kube-api-access-hczqx") pod "b87a1e52-c2d2-4384-97ee-c3cd68e90386" (UID: "b87a1e52-c2d2-4384-97ee-c3cd68e90386"). InnerVolumeSpecName "kube-api-access-hczqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.887024 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "b87a1e52-c2d2-4384-97ee-c3cd68e90386" (UID: "b87a1e52-c2d2-4384-97ee-c3cd68e90386"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.932032 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-inventory" (OuterVolumeSpecName: "inventory") pod "b87a1e52-c2d2-4384-97ee-c3cd68e90386" (UID: "b87a1e52-c2d2-4384-97ee-c3cd68e90386"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.932528 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "b87a1e52-c2d2-4384-97ee-c3cd68e90386" (UID: "b87a1e52-c2d2-4384-97ee-c3cd68e90386"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.934214 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b87a1e52-c2d2-4384-97ee-c3cd68e90386" (UID: "b87a1e52-c2d2-4384-97ee-c3cd68e90386"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.943478 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "b87a1e52-c2d2-4384-97ee-c3cd68e90386" (UID: "b87a1e52-c2d2-4384-97ee-c3cd68e90386"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.947482 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "b87a1e52-c2d2-4384-97ee-c3cd68e90386" (UID: "b87a1e52-c2d2-4384-97ee-c3cd68e90386"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.973980 4934 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.974011 4934 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.974021 4934 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.974030 4934 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.974042 4934 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-inventory\") on node \"crc\" DevicePath \"\"" Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.974053 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hczqx\" (UniqueName: \"kubernetes.io/projected/b87a1e52-c2d2-4384-97ee-c3cd68e90386-kube-api-access-hczqx\") on node \"crc\" DevicePath \"\"" Dec 27 08:34:44 crc kubenswrapper[4934]: I1227 08:34:44.974068 4934 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/b87a1e52-c2d2-4384-97ee-c3cd68e90386-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.260490 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" event={"ID":"b87a1e52-c2d2-4384-97ee-c3cd68e90386","Type":"ContainerDied","Data":"c0c1ad6c580ad447cd2b062b7e8aa03701b0252c65fbb50cf72d424db89ac31f"} Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.260747 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0c1ad6c580ad447cd2b062b7e8aa03701b0252c65fbb50cf72d424db89ac31f" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.260870 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.360944 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw"] Dec 27 08:34:45 crc kubenswrapper[4934]: E1227 08:34:45.361602 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b87a1e52-c2d2-4384-97ee-c3cd68e90386" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.361622 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b87a1e52-c2d2-4384-97ee-c3cd68e90386" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.361861 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="b87a1e52-c2d2-4384-97ee-c3cd68e90386" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.362740 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.366179 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-ipmi-config-data" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.366467 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.367542 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-98ddn" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.368018 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.368158 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.371764 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw"] Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.392762 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.392894 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.392952 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.393058 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.393779 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.395260 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh5tw\" (UniqueName: \"kubernetes.io/projected/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-kube-api-access-dh5tw\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.395320 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.497388 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh5tw\" (UniqueName: \"kubernetes.io/projected/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-kube-api-access-dh5tw\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.497474 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.497593 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.497684 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.497727 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.497816 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.497871 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.501748 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.502765 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.502773 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.502828 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.503364 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.504028 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.514396 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh5tw\" (UniqueName: \"kubernetes.io/projected/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-kube-api-access-dh5tw\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:45 crc kubenswrapper[4934]: I1227 08:34:45.691320 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:34:46 crc kubenswrapper[4934]: I1227 08:34:46.253038 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw"] Dec 27 08:34:46 crc kubenswrapper[4934]: I1227 08:34:46.257509 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 27 08:34:46 crc kubenswrapper[4934]: I1227 08:34:46.277533 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" event={"ID":"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf","Type":"ContainerStarted","Data":"e2b6731b250581c1214dfe35244cf536a95cf9524813db2786f121c044926d45"} Dec 27 08:34:49 crc kubenswrapper[4934]: I1227 08:34:49.313304 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" event={"ID":"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf","Type":"ContainerStarted","Data":"251ecb5a32ac09beb95e1f0a282cb5aa1b3bcf9802c9d085d071c170f0bf46ce"} Dec 27 08:34:49 crc kubenswrapper[4934]: I1227 08:34:49.336702 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" podStartSLOduration=2.137994152 podStartE2EDuration="4.336683476s" podCreationTimestamp="2025-12-27 08:34:45 +0000 UTC" firstStartedPulling="2025-12-27 08:34:46.257219798 +0000 UTC m=+3147.077660392" lastFinishedPulling="2025-12-27 08:34:48.455909122 +0000 UTC m=+3149.276349716" observedRunningTime="2025-12-27 08:34:49.333014626 +0000 UTC m=+3150.153455220" watchObservedRunningTime="2025-12-27 08:34:49.336683476 +0000 UTC m=+3150.157124070" Dec 27 08:34:54 crc kubenswrapper[4934]: I1227 08:34:54.468539 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:34:54 crc kubenswrapper[4934]: E1227 08:34:54.469284 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:35:09 crc kubenswrapper[4934]: I1227 08:35:09.479366 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:35:09 crc kubenswrapper[4934]: E1227 08:35:09.480490 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:35:24 crc kubenswrapper[4934]: I1227 08:35:24.467765 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:35:24 crc kubenswrapper[4934]: E1227 08:35:24.468748 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:35:38 crc kubenswrapper[4934]: I1227 08:35:38.466936 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:35:38 crc kubenswrapper[4934]: E1227 08:35:38.468169 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:35:53 crc kubenswrapper[4934]: I1227 08:35:53.468502 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:35:53 crc kubenswrapper[4934]: E1227 08:35:53.469365 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:36:05 crc kubenswrapper[4934]: I1227 08:36:05.032628 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qtp9d"] Dec 27 08:36:05 crc kubenswrapper[4934]: I1227 08:36:05.036431 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qtp9d" Dec 27 08:36:05 crc kubenswrapper[4934]: I1227 08:36:05.048290 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qtp9d"] Dec 27 08:36:05 crc kubenswrapper[4934]: I1227 08:36:05.099144 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe-utilities\") pod \"community-operators-qtp9d\" (UID: \"0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe\") " pod="openshift-marketplace/community-operators-qtp9d" Dec 27 08:36:05 crc kubenswrapper[4934]: I1227 08:36:05.099211 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe-catalog-content\") pod \"community-operators-qtp9d\" (UID: \"0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe\") " pod="openshift-marketplace/community-operators-qtp9d" Dec 27 08:36:05 crc kubenswrapper[4934]: I1227 08:36:05.099479 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2td7q\" (UniqueName: \"kubernetes.io/projected/0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe-kube-api-access-2td7q\") pod \"community-operators-qtp9d\" (UID: \"0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe\") " pod="openshift-marketplace/community-operators-qtp9d" Dec 27 08:36:05 crc kubenswrapper[4934]: I1227 08:36:05.202123 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe-utilities\") pod \"community-operators-qtp9d\" (UID: \"0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe\") " pod="openshift-marketplace/community-operators-qtp9d" Dec 27 08:36:05 crc kubenswrapper[4934]: I1227 08:36:05.202192 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe-catalog-content\") pod \"community-operators-qtp9d\" (UID: \"0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe\") " pod="openshift-marketplace/community-operators-qtp9d" Dec 27 08:36:05 crc kubenswrapper[4934]: I1227 08:36:05.202273 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2td7q\" (UniqueName: \"kubernetes.io/projected/0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe-kube-api-access-2td7q\") pod \"community-operators-qtp9d\" (UID: \"0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe\") " pod="openshift-marketplace/community-operators-qtp9d" Dec 27 08:36:05 crc kubenswrapper[4934]: I1227 08:36:05.202617 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe-utilities\") pod \"community-operators-qtp9d\" (UID: \"0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe\") " pod="openshift-marketplace/community-operators-qtp9d" Dec 27 08:36:05 crc kubenswrapper[4934]: I1227 08:36:05.202974 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe-catalog-content\") pod \"community-operators-qtp9d\" (UID: \"0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe\") " pod="openshift-marketplace/community-operators-qtp9d" Dec 27 08:36:05 crc kubenswrapper[4934]: I1227 08:36:05.224596 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2td7q\" (UniqueName: \"kubernetes.io/projected/0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe-kube-api-access-2td7q\") pod \"community-operators-qtp9d\" (UID: \"0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe\") " pod="openshift-marketplace/community-operators-qtp9d" Dec 27 08:36:05 crc kubenswrapper[4934]: I1227 08:36:05.370676 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qtp9d" Dec 27 08:36:05 crc kubenswrapper[4934]: I1227 08:36:05.934203 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qtp9d"] Dec 27 08:36:06 crc kubenswrapper[4934]: I1227 08:36:06.291333 4934 generic.go:334] "Generic (PLEG): container finished" podID="0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe" containerID="db03736569dbe9ecbeea91fb63f10d045630ca975d8e9598569b71aece91f52c" exitCode=0 Dec 27 08:36:06 crc kubenswrapper[4934]: I1227 08:36:06.291418 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qtp9d" event={"ID":"0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe","Type":"ContainerDied","Data":"db03736569dbe9ecbeea91fb63f10d045630ca975d8e9598569b71aece91f52c"} Dec 27 08:36:06 crc kubenswrapper[4934]: I1227 08:36:06.291793 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qtp9d" event={"ID":"0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe","Type":"ContainerStarted","Data":"e4836386fad3db00c91bebed25ea2abf2171eb349887fd992d1836c2708e97ff"} Dec 27 08:36:06 crc kubenswrapper[4934]: I1227 08:36:06.467978 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:36:06 crc kubenswrapper[4934]: E1227 08:36:06.468486 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:36:08 crc kubenswrapper[4934]: I1227 08:36:08.319886 4934 generic.go:334] "Generic (PLEG): container finished" podID="0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe" containerID="f0033a0afb21b48e7868405378667ee4f8c4c969c03b3486d12f3a3ea55f1d05" exitCode=0 Dec 27 08:36:08 crc kubenswrapper[4934]: I1227 08:36:08.319933 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qtp9d" event={"ID":"0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe","Type":"ContainerDied","Data":"f0033a0afb21b48e7868405378667ee4f8c4c969c03b3486d12f3a3ea55f1d05"} Dec 27 08:36:10 crc kubenswrapper[4934]: I1227 08:36:10.343378 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qtp9d" event={"ID":"0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe","Type":"ContainerStarted","Data":"cba0544b339438e2b97f2b2e8d6b725d72d9d73a2cb4b3651a72119f3476cfcf"} Dec 27 08:36:10 crc kubenswrapper[4934]: I1227 08:36:10.378549 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qtp9d" podStartSLOduration=2.443420528 podStartE2EDuration="5.378503905s" podCreationTimestamp="2025-12-27 08:36:05 +0000 UTC" firstStartedPulling="2025-12-27 08:36:06.297243573 +0000 UTC m=+3227.117684167" lastFinishedPulling="2025-12-27 08:36:09.23232695 +0000 UTC m=+3230.052767544" observedRunningTime="2025-12-27 08:36:10.367777282 +0000 UTC m=+3231.188217886" watchObservedRunningTime="2025-12-27 08:36:10.378503905 +0000 UTC m=+3231.198944499" Dec 27 08:36:15 crc kubenswrapper[4934]: I1227 08:36:15.371449 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qtp9d" Dec 27 08:36:15 crc kubenswrapper[4934]: I1227 08:36:15.371780 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qtp9d" Dec 27 08:36:15 crc kubenswrapper[4934]: I1227 08:36:15.447431 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qtp9d" Dec 27 08:36:15 crc kubenswrapper[4934]: I1227 08:36:15.498462 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qtp9d" Dec 27 08:36:15 crc kubenswrapper[4934]: I1227 08:36:15.693624 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qtp9d"] Dec 27 08:36:17 crc kubenswrapper[4934]: I1227 08:36:17.416634 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qtp9d" podUID="0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe" containerName="registry-server" containerID="cri-o://cba0544b339438e2b97f2b2e8d6b725d72d9d73a2cb4b3651a72119f3476cfcf" gracePeriod=2 Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.016032 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qtp9d" Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.162626 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe-utilities\") pod \"0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe\" (UID: \"0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe\") " Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.162739 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe-catalog-content\") pod \"0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe\" (UID: \"0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe\") " Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.163239 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2td7q\" (UniqueName: \"kubernetes.io/projected/0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe-kube-api-access-2td7q\") pod \"0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe\" (UID: \"0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe\") " Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.164026 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe-utilities" (OuterVolumeSpecName: "utilities") pod "0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe" (UID: "0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.164363 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.172017 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe-kube-api-access-2td7q" (OuterVolumeSpecName: "kube-api-access-2td7q") pod "0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe" (UID: "0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe"). InnerVolumeSpecName "kube-api-access-2td7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.227406 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe" (UID: "0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.266465 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2td7q\" (UniqueName: \"kubernetes.io/projected/0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe-kube-api-access-2td7q\") on node \"crc\" DevicePath \"\"" Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.266503 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.432661 4934 generic.go:334] "Generic (PLEG): container finished" podID="0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe" containerID="cba0544b339438e2b97f2b2e8d6b725d72d9d73a2cb4b3651a72119f3476cfcf" exitCode=0 Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.432715 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qtp9d" event={"ID":"0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe","Type":"ContainerDied","Data":"cba0544b339438e2b97f2b2e8d6b725d72d9d73a2cb4b3651a72119f3476cfcf"} Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.432748 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qtp9d" Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.432796 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qtp9d" event={"ID":"0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe","Type":"ContainerDied","Data":"e4836386fad3db00c91bebed25ea2abf2171eb349887fd992d1836c2708e97ff"} Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.432829 4934 scope.go:117] "RemoveContainer" containerID="cba0544b339438e2b97f2b2e8d6b725d72d9d73a2cb4b3651a72119f3476cfcf" Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.490249 4934 scope.go:117] "RemoveContainer" containerID="f0033a0afb21b48e7868405378667ee4f8c4c969c03b3486d12f3a3ea55f1d05" Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.496053 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qtp9d"] Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.517291 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qtp9d"] Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.525281 4934 scope.go:117] "RemoveContainer" containerID="db03736569dbe9ecbeea91fb63f10d045630ca975d8e9598569b71aece91f52c" Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.577377 4934 scope.go:117] "RemoveContainer" containerID="cba0544b339438e2b97f2b2e8d6b725d72d9d73a2cb4b3651a72119f3476cfcf" Dec 27 08:36:18 crc kubenswrapper[4934]: E1227 08:36:18.578310 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cba0544b339438e2b97f2b2e8d6b725d72d9d73a2cb4b3651a72119f3476cfcf\": container with ID starting with cba0544b339438e2b97f2b2e8d6b725d72d9d73a2cb4b3651a72119f3476cfcf not found: ID does not exist" containerID="cba0544b339438e2b97f2b2e8d6b725d72d9d73a2cb4b3651a72119f3476cfcf" Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.578350 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cba0544b339438e2b97f2b2e8d6b725d72d9d73a2cb4b3651a72119f3476cfcf"} err="failed to get container status \"cba0544b339438e2b97f2b2e8d6b725d72d9d73a2cb4b3651a72119f3476cfcf\": rpc error: code = NotFound desc = could not find container \"cba0544b339438e2b97f2b2e8d6b725d72d9d73a2cb4b3651a72119f3476cfcf\": container with ID starting with cba0544b339438e2b97f2b2e8d6b725d72d9d73a2cb4b3651a72119f3476cfcf not found: ID does not exist" Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.578379 4934 scope.go:117] "RemoveContainer" containerID="f0033a0afb21b48e7868405378667ee4f8c4c969c03b3486d12f3a3ea55f1d05" Dec 27 08:36:18 crc kubenswrapper[4934]: E1227 08:36:18.579190 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0033a0afb21b48e7868405378667ee4f8c4c969c03b3486d12f3a3ea55f1d05\": container with ID starting with f0033a0afb21b48e7868405378667ee4f8c4c969c03b3486d12f3a3ea55f1d05 not found: ID does not exist" containerID="f0033a0afb21b48e7868405378667ee4f8c4c969c03b3486d12f3a3ea55f1d05" Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.579954 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0033a0afb21b48e7868405378667ee4f8c4c969c03b3486d12f3a3ea55f1d05"} err="failed to get container status \"f0033a0afb21b48e7868405378667ee4f8c4c969c03b3486d12f3a3ea55f1d05\": rpc error: code = NotFound desc = could not find container \"f0033a0afb21b48e7868405378667ee4f8c4c969c03b3486d12f3a3ea55f1d05\": container with ID starting with f0033a0afb21b48e7868405378667ee4f8c4c969c03b3486d12f3a3ea55f1d05 not found: ID does not exist" Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.580011 4934 scope.go:117] "RemoveContainer" containerID="db03736569dbe9ecbeea91fb63f10d045630ca975d8e9598569b71aece91f52c" Dec 27 08:36:18 crc kubenswrapper[4934]: E1227 08:36:18.580542 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db03736569dbe9ecbeea91fb63f10d045630ca975d8e9598569b71aece91f52c\": container with ID starting with db03736569dbe9ecbeea91fb63f10d045630ca975d8e9598569b71aece91f52c not found: ID does not exist" containerID="db03736569dbe9ecbeea91fb63f10d045630ca975d8e9598569b71aece91f52c" Dec 27 08:36:18 crc kubenswrapper[4934]: I1227 08:36:18.580577 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db03736569dbe9ecbeea91fb63f10d045630ca975d8e9598569b71aece91f52c"} err="failed to get container status \"db03736569dbe9ecbeea91fb63f10d045630ca975d8e9598569b71aece91f52c\": rpc error: code = NotFound desc = could not find container \"db03736569dbe9ecbeea91fb63f10d045630ca975d8e9598569b71aece91f52c\": container with ID starting with db03736569dbe9ecbeea91fb63f10d045630ca975d8e9598569b71aece91f52c not found: ID does not exist" Dec 27 08:36:19 crc kubenswrapper[4934]: I1227 08:36:19.495358 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe" path="/var/lib/kubelet/pods/0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe/volumes" Dec 27 08:36:20 crc kubenswrapper[4934]: I1227 08:36:20.468468 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:36:20 crc kubenswrapper[4934]: E1227 08:36:20.469509 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:36:32 crc kubenswrapper[4934]: I1227 08:36:32.468364 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:36:32 crc kubenswrapper[4934]: E1227 08:36:32.469466 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:36:46 crc kubenswrapper[4934]: I1227 08:36:46.468277 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:36:46 crc kubenswrapper[4934]: I1227 08:36:46.878599 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"2e3d7a11613b1624818a75601c039cc83fb1411488789f6d83381d669b30ac39"} Dec 27 08:37:15 crc kubenswrapper[4934]: I1227 08:37:15.250422 4934 generic.go:334] "Generic (PLEG): container finished" podID="54b601bd-dc4a-49b3-bd37-a6d13ae07cdf" containerID="251ecb5a32ac09beb95e1f0a282cb5aa1b3bcf9802c9d085d071c170f0bf46ce" exitCode=0 Dec 27 08:37:15 crc kubenswrapper[4934]: I1227 08:37:15.250490 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" event={"ID":"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf","Type":"ContainerDied","Data":"251ecb5a32ac09beb95e1f0a282cb5aa1b3bcf9802c9d085d071c170f0bf46ce"} Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.764091 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.886577 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-telemetry-power-monitoring-combined-ca-bundle\") pod \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.886691 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dh5tw\" (UniqueName: \"kubernetes.io/projected/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-kube-api-access-dh5tw\") pod \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.886715 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ceilometer-ipmi-config-data-0\") pod \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.886821 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-inventory\") pod \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.886937 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ssh-key\") pod \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.887030 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ceilometer-ipmi-config-data-2\") pod \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.887063 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ceilometer-ipmi-config-data-1\") pod \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\" (UID: \"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf\") " Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.892295 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "54b601bd-dc4a-49b3-bd37-a6d13ae07cdf" (UID: "54b601bd-dc4a-49b3-bd37-a6d13ae07cdf"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.893003 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-kube-api-access-dh5tw" (OuterVolumeSpecName: "kube-api-access-dh5tw") pod "54b601bd-dc4a-49b3-bd37-a6d13ae07cdf" (UID: "54b601bd-dc4a-49b3-bd37-a6d13ae07cdf"). InnerVolumeSpecName "kube-api-access-dh5tw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.918434 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "54b601bd-dc4a-49b3-bd37-a6d13ae07cdf" (UID: "54b601bd-dc4a-49b3-bd37-a6d13ae07cdf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.921385 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ceilometer-ipmi-config-data-1" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-1") pod "54b601bd-dc4a-49b3-bd37-a6d13ae07cdf" (UID: "54b601bd-dc4a-49b3-bd37-a6d13ae07cdf"). InnerVolumeSpecName "ceilometer-ipmi-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.945389 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ceilometer-ipmi-config-data-2" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-2") pod "54b601bd-dc4a-49b3-bd37-a6d13ae07cdf" (UID: "54b601bd-dc4a-49b3-bd37-a6d13ae07cdf"). InnerVolumeSpecName "ceilometer-ipmi-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.949670 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ceilometer-ipmi-config-data-0" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-0") pod "54b601bd-dc4a-49b3-bd37-a6d13ae07cdf" (UID: "54b601bd-dc4a-49b3-bd37-a6d13ae07cdf"). InnerVolumeSpecName "ceilometer-ipmi-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.949851 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-inventory" (OuterVolumeSpecName: "inventory") pod "54b601bd-dc4a-49b3-bd37-a6d13ae07cdf" (UID: "54b601bd-dc4a-49b3-bd37-a6d13ae07cdf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.989710 4934 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ceilometer-ipmi-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.989752 4934 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ceilometer-ipmi-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.989770 4934 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.989790 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dh5tw\" (UniqueName: \"kubernetes.io/projected/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-kube-api-access-dh5tw\") on node \"crc\" DevicePath \"\"" Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.989803 4934 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ceilometer-ipmi-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.989818 4934 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-inventory\") on node \"crc\" DevicePath \"\"" Dec 27 08:37:16 crc kubenswrapper[4934]: I1227 08:37:16.989829 4934 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54b601bd-dc4a-49b3-bd37-a6d13ae07cdf-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.277105 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" event={"ID":"54b601bd-dc4a-49b3-bd37-a6d13ae07cdf","Type":"ContainerDied","Data":"e2b6731b250581c1214dfe35244cf536a95cf9524813db2786f121c044926d45"} Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.277153 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2b6731b250581c1214dfe35244cf536a95cf9524813db2786f121c044926d45" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.277177 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.419288 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f"] Dec 27 08:37:17 crc kubenswrapper[4934]: E1227 08:37:17.420033 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe" containerName="registry-server" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.420065 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe" containerName="registry-server" Dec 27 08:37:17 crc kubenswrapper[4934]: E1227 08:37:17.420117 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54b601bd-dc4a-49b3-bd37-a6d13ae07cdf" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.420131 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="54b601bd-dc4a-49b3-bd37-a6d13ae07cdf" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 27 08:37:17 crc kubenswrapper[4934]: E1227 08:37:17.420155 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe" containerName="extract-content" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.420169 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe" containerName="extract-content" Dec 27 08:37:17 crc kubenswrapper[4934]: E1227 08:37:17.420224 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe" containerName="extract-utilities" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.420237 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe" containerName="extract-utilities" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.420609 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bff9aa5-c2eb-44d2-a8b4-1fde59ae9ffe" containerName="registry-server" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.420682 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="54b601bd-dc4a-49b3-bd37-a6d13ae07cdf" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.422031 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.424267 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.424364 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-98ddn" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.424268 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-compute-config-data" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.424546 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.424638 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.434383 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f"] Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.501470 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-mjh9f\" (UID: \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.501597 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-mjh9f\" (UID: \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.501665 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-mjh9f\" (UID: \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.501945 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vspmk\" (UniqueName: \"kubernetes.io/projected/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-kube-api-access-vspmk\") pod \"logging-edpm-deployment-openstack-edpm-ipam-mjh9f\" (UID: \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.502279 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-mjh9f\" (UID: \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.603963 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-mjh9f\" (UID: \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.604064 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-mjh9f\" (UID: \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.604200 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-mjh9f\" (UID: \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.604251 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-mjh9f\" (UID: \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.604333 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vspmk\" (UniqueName: \"kubernetes.io/projected/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-kube-api-access-vspmk\") pod \"logging-edpm-deployment-openstack-edpm-ipam-mjh9f\" (UID: \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.609117 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-mjh9f\" (UID: \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.609475 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-mjh9f\" (UID: \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.610240 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-mjh9f\" (UID: \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.612704 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-mjh9f\" (UID: \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.627725 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vspmk\" (UniqueName: \"kubernetes.io/projected/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-kube-api-access-vspmk\") pod \"logging-edpm-deployment-openstack-edpm-ipam-mjh9f\" (UID: \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" Dec 27 08:37:17 crc kubenswrapper[4934]: I1227 08:37:17.758508 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" Dec 27 08:37:18 crc kubenswrapper[4934]: I1227 08:37:18.367130 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f"] Dec 27 08:37:19 crc kubenswrapper[4934]: I1227 08:37:19.305986 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" event={"ID":"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9","Type":"ContainerStarted","Data":"11251fd241054ada3464f45bcfaee3542f907d8961eeedf20b653c40648138e5"} Dec 27 08:37:20 crc kubenswrapper[4934]: I1227 08:37:20.318323 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" event={"ID":"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9","Type":"ContainerStarted","Data":"cd0c5490f0fa8f5018ea2c5073c72b8c9c896e3aeb3917de8de04453d9d6b4c0"} Dec 27 08:37:20 crc kubenswrapper[4934]: I1227 08:37:20.352245 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" podStartSLOduration=2.616845755 podStartE2EDuration="3.352206868s" podCreationTimestamp="2025-12-27 08:37:17 +0000 UTC" firstStartedPulling="2025-12-27 08:37:18.365550226 +0000 UTC m=+3299.185990820" lastFinishedPulling="2025-12-27 08:37:19.100887009 +0000 UTC m=+3299.921351933" observedRunningTime="2025-12-27 08:37:20.336971054 +0000 UTC m=+3301.157411688" watchObservedRunningTime="2025-12-27 08:37:20.352206868 +0000 UTC m=+3301.172647532" Dec 27 08:37:35 crc kubenswrapper[4934]: I1227 08:37:35.548264 4934 generic.go:334] "Generic (PLEG): container finished" podID="4d60d2c6-08ec-4bfc-a066-df42d54ec9f9" containerID="cd0c5490f0fa8f5018ea2c5073c72b8c9c896e3aeb3917de8de04453d9d6b4c0" exitCode=0 Dec 27 08:37:35 crc kubenswrapper[4934]: I1227 08:37:35.548339 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" event={"ID":"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9","Type":"ContainerDied","Data":"cd0c5490f0fa8f5018ea2c5073c72b8c9c896e3aeb3917de8de04453d9d6b4c0"} Dec 27 08:37:37 crc kubenswrapper[4934]: I1227 08:37:37.102391 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" Dec 27 08:37:37 crc kubenswrapper[4934]: I1227 08:37:37.234797 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-inventory\") pod \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\" (UID: \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\") " Dec 27 08:37:37 crc kubenswrapper[4934]: I1227 08:37:37.234871 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-logging-compute-config-data-1\") pod \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\" (UID: \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\") " Dec 27 08:37:37 crc kubenswrapper[4934]: I1227 08:37:37.234980 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-ssh-key\") pod \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\" (UID: \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\") " Dec 27 08:37:37 crc kubenswrapper[4934]: I1227 08:37:37.235770 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-logging-compute-config-data-0\") pod \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\" (UID: \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\") " Dec 27 08:37:37 crc kubenswrapper[4934]: I1227 08:37:37.236021 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vspmk\" (UniqueName: \"kubernetes.io/projected/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-kube-api-access-vspmk\") pod \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\" (UID: \"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9\") " Dec 27 08:37:37 crc kubenswrapper[4934]: I1227 08:37:37.243394 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-kube-api-access-vspmk" (OuterVolumeSpecName: "kube-api-access-vspmk") pod "4d60d2c6-08ec-4bfc-a066-df42d54ec9f9" (UID: "4d60d2c6-08ec-4bfc-a066-df42d54ec9f9"). InnerVolumeSpecName "kube-api-access-vspmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:37:37 crc kubenswrapper[4934]: I1227 08:37:37.269235 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-inventory" (OuterVolumeSpecName: "inventory") pod "4d60d2c6-08ec-4bfc-a066-df42d54ec9f9" (UID: "4d60d2c6-08ec-4bfc-a066-df42d54ec9f9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:37:37 crc kubenswrapper[4934]: I1227 08:37:37.271226 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-logging-compute-config-data-0" (OuterVolumeSpecName: "logging-compute-config-data-0") pod "4d60d2c6-08ec-4bfc-a066-df42d54ec9f9" (UID: "4d60d2c6-08ec-4bfc-a066-df42d54ec9f9"). InnerVolumeSpecName "logging-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:37:37 crc kubenswrapper[4934]: I1227 08:37:37.275384 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4d60d2c6-08ec-4bfc-a066-df42d54ec9f9" (UID: "4d60d2c6-08ec-4bfc-a066-df42d54ec9f9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:37:37 crc kubenswrapper[4934]: I1227 08:37:37.291844 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-logging-compute-config-data-1" (OuterVolumeSpecName: "logging-compute-config-data-1") pod "4d60d2c6-08ec-4bfc-a066-df42d54ec9f9" (UID: "4d60d2c6-08ec-4bfc-a066-df42d54ec9f9"). InnerVolumeSpecName "logging-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:37:37 crc kubenswrapper[4934]: I1227 08:37:37.339631 4934 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-inventory\") on node \"crc\" DevicePath \"\"" Dec 27 08:37:37 crc kubenswrapper[4934]: I1227 08:37:37.339665 4934 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-logging-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 27 08:37:37 crc kubenswrapper[4934]: I1227 08:37:37.339680 4934 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 27 08:37:37 crc kubenswrapper[4934]: I1227 08:37:37.339692 4934 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-logging-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 27 08:37:37 crc kubenswrapper[4934]: I1227 08:37:37.339703 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vspmk\" (UniqueName: \"kubernetes.io/projected/4d60d2c6-08ec-4bfc-a066-df42d54ec9f9-kube-api-access-vspmk\") on node \"crc\" DevicePath \"\"" Dec 27 08:37:37 crc kubenswrapper[4934]: I1227 08:37:37.572606 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" event={"ID":"4d60d2c6-08ec-4bfc-a066-df42d54ec9f9","Type":"ContainerDied","Data":"11251fd241054ada3464f45bcfaee3542f907d8961eeedf20b653c40648138e5"} Dec 27 08:37:37 crc kubenswrapper[4934]: I1227 08:37:37.572638 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-mjh9f" Dec 27 08:37:37 crc kubenswrapper[4934]: I1227 08:37:37.572651 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11251fd241054ada3464f45bcfaee3542f907d8961eeedf20b653c40648138e5" Dec 27 08:38:16 crc kubenswrapper[4934]: E1227 08:38:16.618655 4934 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.13:33216->38.102.83.13:43999: write tcp 38.102.83.13:33216->38.102.83.13:43999: write: broken pipe Dec 27 08:39:15 crc kubenswrapper[4934]: I1227 08:39:15.329981 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:39:15 crc kubenswrapper[4934]: I1227 08:39:15.330637 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:39:45 crc kubenswrapper[4934]: I1227 08:39:45.330566 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:39:45 crc kubenswrapper[4934]: I1227 08:39:45.331260 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:40:02 crc kubenswrapper[4934]: I1227 08:40:02.494524 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d4j7v"] Dec 27 08:40:02 crc kubenswrapper[4934]: E1227 08:40:02.495744 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d60d2c6-08ec-4bfc-a066-df42d54ec9f9" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 27 08:40:02 crc kubenswrapper[4934]: I1227 08:40:02.495758 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d60d2c6-08ec-4bfc-a066-df42d54ec9f9" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 27 08:40:02 crc kubenswrapper[4934]: I1227 08:40:02.496020 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d60d2c6-08ec-4bfc-a066-df42d54ec9f9" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 27 08:40:02 crc kubenswrapper[4934]: I1227 08:40:02.497753 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d4j7v" Dec 27 08:40:02 crc kubenswrapper[4934]: I1227 08:40:02.508140 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d4j7v"] Dec 27 08:40:02 crc kubenswrapper[4934]: I1227 08:40:02.624916 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8llkb\" (UniqueName: \"kubernetes.io/projected/9bf6162c-720e-4e1c-80f6-5820304fde5b-kube-api-access-8llkb\") pod \"certified-operators-d4j7v\" (UID: \"9bf6162c-720e-4e1c-80f6-5820304fde5b\") " pod="openshift-marketplace/certified-operators-d4j7v" Dec 27 08:40:02 crc kubenswrapper[4934]: I1227 08:40:02.625030 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bf6162c-720e-4e1c-80f6-5820304fde5b-catalog-content\") pod \"certified-operators-d4j7v\" (UID: \"9bf6162c-720e-4e1c-80f6-5820304fde5b\") " pod="openshift-marketplace/certified-operators-d4j7v" Dec 27 08:40:02 crc kubenswrapper[4934]: I1227 08:40:02.625198 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bf6162c-720e-4e1c-80f6-5820304fde5b-utilities\") pod \"certified-operators-d4j7v\" (UID: \"9bf6162c-720e-4e1c-80f6-5820304fde5b\") " pod="openshift-marketplace/certified-operators-d4j7v" Dec 27 08:40:02 crc kubenswrapper[4934]: I1227 08:40:02.726994 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8llkb\" (UniqueName: \"kubernetes.io/projected/9bf6162c-720e-4e1c-80f6-5820304fde5b-kube-api-access-8llkb\") pod \"certified-operators-d4j7v\" (UID: \"9bf6162c-720e-4e1c-80f6-5820304fde5b\") " pod="openshift-marketplace/certified-operators-d4j7v" Dec 27 08:40:02 crc kubenswrapper[4934]: I1227 08:40:02.727062 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bf6162c-720e-4e1c-80f6-5820304fde5b-catalog-content\") pod \"certified-operators-d4j7v\" (UID: \"9bf6162c-720e-4e1c-80f6-5820304fde5b\") " pod="openshift-marketplace/certified-operators-d4j7v" Dec 27 08:40:02 crc kubenswrapper[4934]: I1227 08:40:02.727215 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bf6162c-720e-4e1c-80f6-5820304fde5b-utilities\") pod \"certified-operators-d4j7v\" (UID: \"9bf6162c-720e-4e1c-80f6-5820304fde5b\") " pod="openshift-marketplace/certified-operators-d4j7v" Dec 27 08:40:02 crc kubenswrapper[4934]: I1227 08:40:02.727724 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bf6162c-720e-4e1c-80f6-5820304fde5b-utilities\") pod \"certified-operators-d4j7v\" (UID: \"9bf6162c-720e-4e1c-80f6-5820304fde5b\") " pod="openshift-marketplace/certified-operators-d4j7v" Dec 27 08:40:02 crc kubenswrapper[4934]: I1227 08:40:02.727764 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bf6162c-720e-4e1c-80f6-5820304fde5b-catalog-content\") pod \"certified-operators-d4j7v\" (UID: \"9bf6162c-720e-4e1c-80f6-5820304fde5b\") " pod="openshift-marketplace/certified-operators-d4j7v" Dec 27 08:40:02 crc kubenswrapper[4934]: I1227 08:40:02.761075 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8llkb\" (UniqueName: \"kubernetes.io/projected/9bf6162c-720e-4e1c-80f6-5820304fde5b-kube-api-access-8llkb\") pod \"certified-operators-d4j7v\" (UID: \"9bf6162c-720e-4e1c-80f6-5820304fde5b\") " pod="openshift-marketplace/certified-operators-d4j7v" Dec 27 08:40:02 crc kubenswrapper[4934]: I1227 08:40:02.836877 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d4j7v" Dec 27 08:40:03 crc kubenswrapper[4934]: I1227 08:40:03.423362 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d4j7v"] Dec 27 08:40:04 crc kubenswrapper[4934]: I1227 08:40:04.352810 4934 generic.go:334] "Generic (PLEG): container finished" podID="9bf6162c-720e-4e1c-80f6-5820304fde5b" containerID="ddecd787e36b34f17f29527ce76ff04e44fb3562058f4e2ad9eb3389e3dbbf32" exitCode=0 Dec 27 08:40:04 crc kubenswrapper[4934]: I1227 08:40:04.352871 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4j7v" event={"ID":"9bf6162c-720e-4e1c-80f6-5820304fde5b","Type":"ContainerDied","Data":"ddecd787e36b34f17f29527ce76ff04e44fb3562058f4e2ad9eb3389e3dbbf32"} Dec 27 08:40:04 crc kubenswrapper[4934]: I1227 08:40:04.353171 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4j7v" event={"ID":"9bf6162c-720e-4e1c-80f6-5820304fde5b","Type":"ContainerStarted","Data":"3ad34f35b33e77b4b5a7a274f0de2304319e70ad803121002dde4fce135c92d7"} Dec 27 08:40:04 crc kubenswrapper[4934]: I1227 08:40:04.355905 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 27 08:40:06 crc kubenswrapper[4934]: I1227 08:40:06.389519 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4j7v" event={"ID":"9bf6162c-720e-4e1c-80f6-5820304fde5b","Type":"ContainerStarted","Data":"68d9d94bc7ba707da9cea416dd9fcb21a49f4d331149636f6c5a3f52c27e2a34"} Dec 27 08:40:07 crc kubenswrapper[4934]: I1227 08:40:07.406158 4934 generic.go:334] "Generic (PLEG): container finished" podID="9bf6162c-720e-4e1c-80f6-5820304fde5b" containerID="68d9d94bc7ba707da9cea416dd9fcb21a49f4d331149636f6c5a3f52c27e2a34" exitCode=0 Dec 27 08:40:07 crc kubenswrapper[4934]: I1227 08:40:07.406209 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4j7v" event={"ID":"9bf6162c-720e-4e1c-80f6-5820304fde5b","Type":"ContainerDied","Data":"68d9d94bc7ba707da9cea416dd9fcb21a49f4d331149636f6c5a3f52c27e2a34"} Dec 27 08:40:08 crc kubenswrapper[4934]: I1227 08:40:08.420214 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4j7v" event={"ID":"9bf6162c-720e-4e1c-80f6-5820304fde5b","Type":"ContainerStarted","Data":"9b775fb41beaba02c2cff4d67f361fafd97612cb9702f5c16f3a25cc6d74773a"} Dec 27 08:40:08 crc kubenswrapper[4934]: I1227 08:40:08.447477 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d4j7v" podStartSLOduration=2.784011195 podStartE2EDuration="6.447460746s" podCreationTimestamp="2025-12-27 08:40:02 +0000 UTC" firstStartedPulling="2025-12-27 08:40:04.355614612 +0000 UTC m=+3465.176055216" lastFinishedPulling="2025-12-27 08:40:08.019064173 +0000 UTC m=+3468.839504767" observedRunningTime="2025-12-27 08:40:08.442116445 +0000 UTC m=+3469.262557049" watchObservedRunningTime="2025-12-27 08:40:08.447460746 +0000 UTC m=+3469.267901350" Dec 27 08:40:12 crc kubenswrapper[4934]: I1227 08:40:12.837675 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d4j7v" Dec 27 08:40:12 crc kubenswrapper[4934]: I1227 08:40:12.840967 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d4j7v" Dec 27 08:40:12 crc kubenswrapper[4934]: I1227 08:40:12.912529 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d4j7v" Dec 27 08:40:13 crc kubenswrapper[4934]: I1227 08:40:13.555801 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d4j7v" Dec 27 08:40:13 crc kubenswrapper[4934]: I1227 08:40:13.628741 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d4j7v"] Dec 27 08:40:15 crc kubenswrapper[4934]: I1227 08:40:15.330335 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:40:15 crc kubenswrapper[4934]: I1227 08:40:15.330801 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:40:15 crc kubenswrapper[4934]: I1227 08:40:15.330885 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 08:40:15 crc kubenswrapper[4934]: I1227 08:40:15.332307 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2e3d7a11613b1624818a75601c039cc83fb1411488789f6d83381d669b30ac39"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 08:40:15 crc kubenswrapper[4934]: I1227 08:40:15.332455 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://2e3d7a11613b1624818a75601c039cc83fb1411488789f6d83381d669b30ac39" gracePeriod=600 Dec 27 08:40:15 crc kubenswrapper[4934]: I1227 08:40:15.525418 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="2e3d7a11613b1624818a75601c039cc83fb1411488789f6d83381d669b30ac39" exitCode=0 Dec 27 08:40:15 crc kubenswrapper[4934]: I1227 08:40:15.525696 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-d4j7v" podUID="9bf6162c-720e-4e1c-80f6-5820304fde5b" containerName="registry-server" containerID="cri-o://9b775fb41beaba02c2cff4d67f361fafd97612cb9702f5c16f3a25cc6d74773a" gracePeriod=2 Dec 27 08:40:15 crc kubenswrapper[4934]: I1227 08:40:15.526808 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"2e3d7a11613b1624818a75601c039cc83fb1411488789f6d83381d669b30ac39"} Dec 27 08:40:15 crc kubenswrapper[4934]: I1227 08:40:15.526922 4934 scope.go:117] "RemoveContainer" containerID="5572690f2e8e5b8ed6a519ed048d4d325844de4fdadc0f7920ef5381c9b9a19b" Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.135917 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d4j7v" Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.231821 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bf6162c-720e-4e1c-80f6-5820304fde5b-catalog-content\") pod \"9bf6162c-720e-4e1c-80f6-5820304fde5b\" (UID: \"9bf6162c-720e-4e1c-80f6-5820304fde5b\") " Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.232053 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8llkb\" (UniqueName: \"kubernetes.io/projected/9bf6162c-720e-4e1c-80f6-5820304fde5b-kube-api-access-8llkb\") pod \"9bf6162c-720e-4e1c-80f6-5820304fde5b\" (UID: \"9bf6162c-720e-4e1c-80f6-5820304fde5b\") " Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.232122 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bf6162c-720e-4e1c-80f6-5820304fde5b-utilities\") pod \"9bf6162c-720e-4e1c-80f6-5820304fde5b\" (UID: \"9bf6162c-720e-4e1c-80f6-5820304fde5b\") " Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.235129 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bf6162c-720e-4e1c-80f6-5820304fde5b-utilities" (OuterVolumeSpecName: "utilities") pod "9bf6162c-720e-4e1c-80f6-5820304fde5b" (UID: "9bf6162c-720e-4e1c-80f6-5820304fde5b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.238765 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bf6162c-720e-4e1c-80f6-5820304fde5b-kube-api-access-8llkb" (OuterVolumeSpecName: "kube-api-access-8llkb") pod "9bf6162c-720e-4e1c-80f6-5820304fde5b" (UID: "9bf6162c-720e-4e1c-80f6-5820304fde5b"). InnerVolumeSpecName "kube-api-access-8llkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.283463 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bf6162c-720e-4e1c-80f6-5820304fde5b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9bf6162c-720e-4e1c-80f6-5820304fde5b" (UID: "9bf6162c-720e-4e1c-80f6-5820304fde5b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.336309 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bf6162c-720e-4e1c-80f6-5820304fde5b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.336363 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8llkb\" (UniqueName: \"kubernetes.io/projected/9bf6162c-720e-4e1c-80f6-5820304fde5b-kube-api-access-8llkb\") on node \"crc\" DevicePath \"\"" Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.336385 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bf6162c-720e-4e1c-80f6-5820304fde5b-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.540926 4934 generic.go:334] "Generic (PLEG): container finished" podID="9bf6162c-720e-4e1c-80f6-5820304fde5b" containerID="9b775fb41beaba02c2cff4d67f361fafd97612cb9702f5c16f3a25cc6d74773a" exitCode=0 Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.541013 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d4j7v" Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.541027 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4j7v" event={"ID":"9bf6162c-720e-4e1c-80f6-5820304fde5b","Type":"ContainerDied","Data":"9b775fb41beaba02c2cff4d67f361fafd97612cb9702f5c16f3a25cc6d74773a"} Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.541069 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d4j7v" event={"ID":"9bf6162c-720e-4e1c-80f6-5820304fde5b","Type":"ContainerDied","Data":"3ad34f35b33e77b4b5a7a274f0de2304319e70ad803121002dde4fce135c92d7"} Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.541120 4934 scope.go:117] "RemoveContainer" containerID="9b775fb41beaba02c2cff4d67f361fafd97612cb9702f5c16f3a25cc6d74773a" Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.545355 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754"} Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.593710 4934 scope.go:117] "RemoveContainer" containerID="68d9d94bc7ba707da9cea416dd9fcb21a49f4d331149636f6c5a3f52c27e2a34" Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.610718 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d4j7v"] Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.620488 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-d4j7v"] Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.634775 4934 scope.go:117] "RemoveContainer" containerID="ddecd787e36b34f17f29527ce76ff04e44fb3562058f4e2ad9eb3389e3dbbf32" Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.676887 4934 scope.go:117] "RemoveContainer" containerID="9b775fb41beaba02c2cff4d67f361fafd97612cb9702f5c16f3a25cc6d74773a" Dec 27 08:40:16 crc kubenswrapper[4934]: E1227 08:40:16.677454 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b775fb41beaba02c2cff4d67f361fafd97612cb9702f5c16f3a25cc6d74773a\": container with ID starting with 9b775fb41beaba02c2cff4d67f361fafd97612cb9702f5c16f3a25cc6d74773a not found: ID does not exist" containerID="9b775fb41beaba02c2cff4d67f361fafd97612cb9702f5c16f3a25cc6d74773a" Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.677504 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b775fb41beaba02c2cff4d67f361fafd97612cb9702f5c16f3a25cc6d74773a"} err="failed to get container status \"9b775fb41beaba02c2cff4d67f361fafd97612cb9702f5c16f3a25cc6d74773a\": rpc error: code = NotFound desc = could not find container \"9b775fb41beaba02c2cff4d67f361fafd97612cb9702f5c16f3a25cc6d74773a\": container with ID starting with 9b775fb41beaba02c2cff4d67f361fafd97612cb9702f5c16f3a25cc6d74773a not found: ID does not exist" Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.677541 4934 scope.go:117] "RemoveContainer" containerID="68d9d94bc7ba707da9cea416dd9fcb21a49f4d331149636f6c5a3f52c27e2a34" Dec 27 08:40:16 crc kubenswrapper[4934]: E1227 08:40:16.677878 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68d9d94bc7ba707da9cea416dd9fcb21a49f4d331149636f6c5a3f52c27e2a34\": container with ID starting with 68d9d94bc7ba707da9cea416dd9fcb21a49f4d331149636f6c5a3f52c27e2a34 not found: ID does not exist" containerID="68d9d94bc7ba707da9cea416dd9fcb21a49f4d331149636f6c5a3f52c27e2a34" Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.677917 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68d9d94bc7ba707da9cea416dd9fcb21a49f4d331149636f6c5a3f52c27e2a34"} err="failed to get container status \"68d9d94bc7ba707da9cea416dd9fcb21a49f4d331149636f6c5a3f52c27e2a34\": rpc error: code = NotFound desc = could not find container \"68d9d94bc7ba707da9cea416dd9fcb21a49f4d331149636f6c5a3f52c27e2a34\": container with ID starting with 68d9d94bc7ba707da9cea416dd9fcb21a49f4d331149636f6c5a3f52c27e2a34 not found: ID does not exist" Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.677942 4934 scope.go:117] "RemoveContainer" containerID="ddecd787e36b34f17f29527ce76ff04e44fb3562058f4e2ad9eb3389e3dbbf32" Dec 27 08:40:16 crc kubenswrapper[4934]: E1227 08:40:16.678379 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddecd787e36b34f17f29527ce76ff04e44fb3562058f4e2ad9eb3389e3dbbf32\": container with ID starting with ddecd787e36b34f17f29527ce76ff04e44fb3562058f4e2ad9eb3389e3dbbf32 not found: ID does not exist" containerID="ddecd787e36b34f17f29527ce76ff04e44fb3562058f4e2ad9eb3389e3dbbf32" Dec 27 08:40:16 crc kubenswrapper[4934]: I1227 08:40:16.678581 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddecd787e36b34f17f29527ce76ff04e44fb3562058f4e2ad9eb3389e3dbbf32"} err="failed to get container status \"ddecd787e36b34f17f29527ce76ff04e44fb3562058f4e2ad9eb3389e3dbbf32\": rpc error: code = NotFound desc = could not find container \"ddecd787e36b34f17f29527ce76ff04e44fb3562058f4e2ad9eb3389e3dbbf32\": container with ID starting with ddecd787e36b34f17f29527ce76ff04e44fb3562058f4e2ad9eb3389e3dbbf32 not found: ID does not exist" Dec 27 08:40:17 crc kubenswrapper[4934]: I1227 08:40:17.483421 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bf6162c-720e-4e1c-80f6-5820304fde5b" path="/var/lib/kubelet/pods/9bf6162c-720e-4e1c-80f6-5820304fde5b/volumes" Dec 27 08:40:23 crc kubenswrapper[4934]: I1227 08:40:23.336345 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zsrlq"] Dec 27 08:40:23 crc kubenswrapper[4934]: E1227 08:40:23.338004 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bf6162c-720e-4e1c-80f6-5820304fde5b" containerName="extract-content" Dec 27 08:40:23 crc kubenswrapper[4934]: I1227 08:40:23.338056 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bf6162c-720e-4e1c-80f6-5820304fde5b" containerName="extract-content" Dec 27 08:40:23 crc kubenswrapper[4934]: E1227 08:40:23.338139 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bf6162c-720e-4e1c-80f6-5820304fde5b" containerName="extract-utilities" Dec 27 08:40:23 crc kubenswrapper[4934]: I1227 08:40:23.338148 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bf6162c-720e-4e1c-80f6-5820304fde5b" containerName="extract-utilities" Dec 27 08:40:23 crc kubenswrapper[4934]: E1227 08:40:23.338190 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bf6162c-720e-4e1c-80f6-5820304fde5b" containerName="registry-server" Dec 27 08:40:23 crc kubenswrapper[4934]: I1227 08:40:23.338196 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bf6162c-720e-4e1c-80f6-5820304fde5b" containerName="registry-server" Dec 27 08:40:23 crc kubenswrapper[4934]: I1227 08:40:23.338486 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bf6162c-720e-4e1c-80f6-5820304fde5b" containerName="registry-server" Dec 27 08:40:23 crc kubenswrapper[4934]: I1227 08:40:23.341355 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zsrlq" Dec 27 08:40:23 crc kubenswrapper[4934]: I1227 08:40:23.361964 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zsrlq"] Dec 27 08:40:23 crc kubenswrapper[4934]: I1227 08:40:23.440392 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9155f175-88a3-462a-881a-a739684f8bab-utilities\") pod \"redhat-marketplace-zsrlq\" (UID: \"9155f175-88a3-462a-881a-a739684f8bab\") " pod="openshift-marketplace/redhat-marketplace-zsrlq" Dec 27 08:40:23 crc kubenswrapper[4934]: I1227 08:40:23.440768 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9155f175-88a3-462a-881a-a739684f8bab-catalog-content\") pod \"redhat-marketplace-zsrlq\" (UID: \"9155f175-88a3-462a-881a-a739684f8bab\") " pod="openshift-marketplace/redhat-marketplace-zsrlq" Dec 27 08:40:23 crc kubenswrapper[4934]: I1227 08:40:23.440959 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2dxb\" (UniqueName: \"kubernetes.io/projected/9155f175-88a3-462a-881a-a739684f8bab-kube-api-access-m2dxb\") pod \"redhat-marketplace-zsrlq\" (UID: \"9155f175-88a3-462a-881a-a739684f8bab\") " pod="openshift-marketplace/redhat-marketplace-zsrlq" Dec 27 08:40:23 crc kubenswrapper[4934]: I1227 08:40:23.543532 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9155f175-88a3-462a-881a-a739684f8bab-catalog-content\") pod \"redhat-marketplace-zsrlq\" (UID: \"9155f175-88a3-462a-881a-a739684f8bab\") " pod="openshift-marketplace/redhat-marketplace-zsrlq" Dec 27 08:40:23 crc kubenswrapper[4934]: I1227 08:40:23.544024 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2dxb\" (UniqueName: \"kubernetes.io/projected/9155f175-88a3-462a-881a-a739684f8bab-kube-api-access-m2dxb\") pod \"redhat-marketplace-zsrlq\" (UID: \"9155f175-88a3-462a-881a-a739684f8bab\") " pod="openshift-marketplace/redhat-marketplace-zsrlq" Dec 27 08:40:23 crc kubenswrapper[4934]: I1227 08:40:23.544299 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9155f175-88a3-462a-881a-a739684f8bab-utilities\") pod \"redhat-marketplace-zsrlq\" (UID: \"9155f175-88a3-462a-881a-a739684f8bab\") " pod="openshift-marketplace/redhat-marketplace-zsrlq" Dec 27 08:40:23 crc kubenswrapper[4934]: I1227 08:40:23.544527 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9155f175-88a3-462a-881a-a739684f8bab-catalog-content\") pod \"redhat-marketplace-zsrlq\" (UID: \"9155f175-88a3-462a-881a-a739684f8bab\") " pod="openshift-marketplace/redhat-marketplace-zsrlq" Dec 27 08:40:23 crc kubenswrapper[4934]: I1227 08:40:23.544888 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9155f175-88a3-462a-881a-a739684f8bab-utilities\") pod \"redhat-marketplace-zsrlq\" (UID: \"9155f175-88a3-462a-881a-a739684f8bab\") " pod="openshift-marketplace/redhat-marketplace-zsrlq" Dec 27 08:40:23 crc kubenswrapper[4934]: I1227 08:40:23.564309 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2dxb\" (UniqueName: \"kubernetes.io/projected/9155f175-88a3-462a-881a-a739684f8bab-kube-api-access-m2dxb\") pod \"redhat-marketplace-zsrlq\" (UID: \"9155f175-88a3-462a-881a-a739684f8bab\") " pod="openshift-marketplace/redhat-marketplace-zsrlq" Dec 27 08:40:23 crc kubenswrapper[4934]: I1227 08:40:23.662762 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zsrlq" Dec 27 08:40:24 crc kubenswrapper[4934]: I1227 08:40:24.167898 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zsrlq"] Dec 27 08:40:24 crc kubenswrapper[4934]: I1227 08:40:24.677912 4934 generic.go:334] "Generic (PLEG): container finished" podID="9155f175-88a3-462a-881a-a739684f8bab" containerID="3c9af1b0f051b44d51703058c3da580a91f6c3841feadbf718e96ad04203da9b" exitCode=0 Dec 27 08:40:24 crc kubenswrapper[4934]: I1227 08:40:24.678394 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zsrlq" event={"ID":"9155f175-88a3-462a-881a-a739684f8bab","Type":"ContainerDied","Data":"3c9af1b0f051b44d51703058c3da580a91f6c3841feadbf718e96ad04203da9b"} Dec 27 08:40:24 crc kubenswrapper[4934]: I1227 08:40:24.678449 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zsrlq" event={"ID":"9155f175-88a3-462a-881a-a739684f8bab","Type":"ContainerStarted","Data":"0f6633b9bf7b91dc1c28d512393f67efeaa09d35ed261162a0f2aedbf9e948f7"} Dec 27 08:40:26 crc kubenswrapper[4934]: I1227 08:40:26.714632 4934 generic.go:334] "Generic (PLEG): container finished" podID="9155f175-88a3-462a-881a-a739684f8bab" containerID="a261401a6fea899e7c1f486ed94415faf6c656bd3d8ff6e9b49ffcf28ac09077" exitCode=0 Dec 27 08:40:26 crc kubenswrapper[4934]: I1227 08:40:26.714711 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zsrlq" event={"ID":"9155f175-88a3-462a-881a-a739684f8bab","Type":"ContainerDied","Data":"a261401a6fea899e7c1f486ed94415faf6c656bd3d8ff6e9b49ffcf28ac09077"} Dec 27 08:40:27 crc kubenswrapper[4934]: I1227 08:40:27.728503 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zsrlq" event={"ID":"9155f175-88a3-462a-881a-a739684f8bab","Type":"ContainerStarted","Data":"9063ef37726d833f8426ff7405e1fa6fed23f0473cfaba26229d0666e8239077"} Dec 27 08:40:27 crc kubenswrapper[4934]: I1227 08:40:27.779121 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zsrlq" podStartSLOduration=2.214975155 podStartE2EDuration="4.779073899s" podCreationTimestamp="2025-12-27 08:40:23 +0000 UTC" firstStartedPulling="2025-12-27 08:40:24.682904792 +0000 UTC m=+3485.503345426" lastFinishedPulling="2025-12-27 08:40:27.247003576 +0000 UTC m=+3488.067444170" observedRunningTime="2025-12-27 08:40:27.757035009 +0000 UTC m=+3488.577475603" watchObservedRunningTime="2025-12-27 08:40:27.779073899 +0000 UTC m=+3488.599514523" Dec 27 08:40:33 crc kubenswrapper[4934]: I1227 08:40:33.663590 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zsrlq" Dec 27 08:40:33 crc kubenswrapper[4934]: I1227 08:40:33.664443 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zsrlq" Dec 27 08:40:33 crc kubenswrapper[4934]: I1227 08:40:33.748004 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zsrlq" Dec 27 08:40:33 crc kubenswrapper[4934]: I1227 08:40:33.896357 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zsrlq" Dec 27 08:40:34 crc kubenswrapper[4934]: I1227 08:40:34.007615 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zsrlq"] Dec 27 08:40:35 crc kubenswrapper[4934]: I1227 08:40:35.851866 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zsrlq" podUID="9155f175-88a3-462a-881a-a739684f8bab" containerName="registry-server" containerID="cri-o://9063ef37726d833f8426ff7405e1fa6fed23f0473cfaba26229d0666e8239077" gracePeriod=2 Dec 27 08:40:36 crc kubenswrapper[4934]: I1227 08:40:36.445462 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zsrlq" Dec 27 08:40:36 crc kubenswrapper[4934]: I1227 08:40:36.511547 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9155f175-88a3-462a-881a-a739684f8bab-catalog-content\") pod \"9155f175-88a3-462a-881a-a739684f8bab\" (UID: \"9155f175-88a3-462a-881a-a739684f8bab\") " Dec 27 08:40:36 crc kubenswrapper[4934]: I1227 08:40:36.511726 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9155f175-88a3-462a-881a-a739684f8bab-utilities\") pod \"9155f175-88a3-462a-881a-a739684f8bab\" (UID: \"9155f175-88a3-462a-881a-a739684f8bab\") " Dec 27 08:40:36 crc kubenswrapper[4934]: I1227 08:40:36.511952 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2dxb\" (UniqueName: \"kubernetes.io/projected/9155f175-88a3-462a-881a-a739684f8bab-kube-api-access-m2dxb\") pod \"9155f175-88a3-462a-881a-a739684f8bab\" (UID: \"9155f175-88a3-462a-881a-a739684f8bab\") " Dec 27 08:40:36 crc kubenswrapper[4934]: I1227 08:40:36.514427 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9155f175-88a3-462a-881a-a739684f8bab-utilities" (OuterVolumeSpecName: "utilities") pod "9155f175-88a3-462a-881a-a739684f8bab" (UID: "9155f175-88a3-462a-881a-a739684f8bab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:40:36 crc kubenswrapper[4934]: I1227 08:40:36.524492 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9155f175-88a3-462a-881a-a739684f8bab-kube-api-access-m2dxb" (OuterVolumeSpecName: "kube-api-access-m2dxb") pod "9155f175-88a3-462a-881a-a739684f8bab" (UID: "9155f175-88a3-462a-881a-a739684f8bab"). InnerVolumeSpecName "kube-api-access-m2dxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:40:36 crc kubenswrapper[4934]: I1227 08:40:36.542015 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9155f175-88a3-462a-881a-a739684f8bab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9155f175-88a3-462a-881a-a739684f8bab" (UID: "9155f175-88a3-462a-881a-a739684f8bab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:40:36 crc kubenswrapper[4934]: I1227 08:40:36.615969 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2dxb\" (UniqueName: \"kubernetes.io/projected/9155f175-88a3-462a-881a-a739684f8bab-kube-api-access-m2dxb\") on node \"crc\" DevicePath \"\"" Dec 27 08:40:36 crc kubenswrapper[4934]: I1227 08:40:36.616208 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9155f175-88a3-462a-881a-a739684f8bab-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 08:40:36 crc kubenswrapper[4934]: I1227 08:40:36.616271 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9155f175-88a3-462a-881a-a739684f8bab-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 08:40:36 crc kubenswrapper[4934]: I1227 08:40:36.867867 4934 generic.go:334] "Generic (PLEG): container finished" podID="9155f175-88a3-462a-881a-a739684f8bab" containerID="9063ef37726d833f8426ff7405e1fa6fed23f0473cfaba26229d0666e8239077" exitCode=0 Dec 27 08:40:36 crc kubenswrapper[4934]: I1227 08:40:36.867929 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zsrlq" Dec 27 08:40:36 crc kubenswrapper[4934]: I1227 08:40:36.867930 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zsrlq" event={"ID":"9155f175-88a3-462a-881a-a739684f8bab","Type":"ContainerDied","Data":"9063ef37726d833f8426ff7405e1fa6fed23f0473cfaba26229d0666e8239077"} Dec 27 08:40:36 crc kubenswrapper[4934]: I1227 08:40:36.867992 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zsrlq" event={"ID":"9155f175-88a3-462a-881a-a739684f8bab","Type":"ContainerDied","Data":"0f6633b9bf7b91dc1c28d512393f67efeaa09d35ed261162a0f2aedbf9e948f7"} Dec 27 08:40:36 crc kubenswrapper[4934]: I1227 08:40:36.868014 4934 scope.go:117] "RemoveContainer" containerID="9063ef37726d833f8426ff7405e1fa6fed23f0473cfaba26229d0666e8239077" Dec 27 08:40:36 crc kubenswrapper[4934]: I1227 08:40:36.911512 4934 scope.go:117] "RemoveContainer" containerID="a261401a6fea899e7c1f486ed94415faf6c656bd3d8ff6e9b49ffcf28ac09077" Dec 27 08:40:36 crc kubenswrapper[4934]: I1227 08:40:36.915609 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zsrlq"] Dec 27 08:40:36 crc kubenswrapper[4934]: I1227 08:40:36.928663 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zsrlq"] Dec 27 08:40:36 crc kubenswrapper[4934]: I1227 08:40:36.945063 4934 scope.go:117] "RemoveContainer" containerID="3c9af1b0f051b44d51703058c3da580a91f6c3841feadbf718e96ad04203da9b" Dec 27 08:40:37 crc kubenswrapper[4934]: I1227 08:40:37.017024 4934 scope.go:117] "RemoveContainer" containerID="9063ef37726d833f8426ff7405e1fa6fed23f0473cfaba26229d0666e8239077" Dec 27 08:40:37 crc kubenswrapper[4934]: E1227 08:40:37.017668 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9063ef37726d833f8426ff7405e1fa6fed23f0473cfaba26229d0666e8239077\": container with ID starting with 9063ef37726d833f8426ff7405e1fa6fed23f0473cfaba26229d0666e8239077 not found: ID does not exist" containerID="9063ef37726d833f8426ff7405e1fa6fed23f0473cfaba26229d0666e8239077" Dec 27 08:40:37 crc kubenswrapper[4934]: I1227 08:40:37.017726 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9063ef37726d833f8426ff7405e1fa6fed23f0473cfaba26229d0666e8239077"} err="failed to get container status \"9063ef37726d833f8426ff7405e1fa6fed23f0473cfaba26229d0666e8239077\": rpc error: code = NotFound desc = could not find container \"9063ef37726d833f8426ff7405e1fa6fed23f0473cfaba26229d0666e8239077\": container with ID starting with 9063ef37726d833f8426ff7405e1fa6fed23f0473cfaba26229d0666e8239077 not found: ID does not exist" Dec 27 08:40:37 crc kubenswrapper[4934]: I1227 08:40:37.017765 4934 scope.go:117] "RemoveContainer" containerID="a261401a6fea899e7c1f486ed94415faf6c656bd3d8ff6e9b49ffcf28ac09077" Dec 27 08:40:37 crc kubenswrapper[4934]: E1227 08:40:37.018427 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a261401a6fea899e7c1f486ed94415faf6c656bd3d8ff6e9b49ffcf28ac09077\": container with ID starting with a261401a6fea899e7c1f486ed94415faf6c656bd3d8ff6e9b49ffcf28ac09077 not found: ID does not exist" containerID="a261401a6fea899e7c1f486ed94415faf6c656bd3d8ff6e9b49ffcf28ac09077" Dec 27 08:40:37 crc kubenswrapper[4934]: I1227 08:40:37.018489 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a261401a6fea899e7c1f486ed94415faf6c656bd3d8ff6e9b49ffcf28ac09077"} err="failed to get container status \"a261401a6fea899e7c1f486ed94415faf6c656bd3d8ff6e9b49ffcf28ac09077\": rpc error: code = NotFound desc = could not find container \"a261401a6fea899e7c1f486ed94415faf6c656bd3d8ff6e9b49ffcf28ac09077\": container with ID starting with a261401a6fea899e7c1f486ed94415faf6c656bd3d8ff6e9b49ffcf28ac09077 not found: ID does not exist" Dec 27 08:40:37 crc kubenswrapper[4934]: I1227 08:40:37.018530 4934 scope.go:117] "RemoveContainer" containerID="3c9af1b0f051b44d51703058c3da580a91f6c3841feadbf718e96ad04203da9b" Dec 27 08:40:37 crc kubenswrapper[4934]: E1227 08:40:37.019007 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c9af1b0f051b44d51703058c3da580a91f6c3841feadbf718e96ad04203da9b\": container with ID starting with 3c9af1b0f051b44d51703058c3da580a91f6c3841feadbf718e96ad04203da9b not found: ID does not exist" containerID="3c9af1b0f051b44d51703058c3da580a91f6c3841feadbf718e96ad04203da9b" Dec 27 08:40:37 crc kubenswrapper[4934]: I1227 08:40:37.019242 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c9af1b0f051b44d51703058c3da580a91f6c3841feadbf718e96ad04203da9b"} err="failed to get container status \"3c9af1b0f051b44d51703058c3da580a91f6c3841feadbf718e96ad04203da9b\": rpc error: code = NotFound desc = could not find container \"3c9af1b0f051b44d51703058c3da580a91f6c3841feadbf718e96ad04203da9b\": container with ID starting with 3c9af1b0f051b44d51703058c3da580a91f6c3841feadbf718e96ad04203da9b not found: ID does not exist" Dec 27 08:40:37 crc kubenswrapper[4934]: I1227 08:40:37.480320 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9155f175-88a3-462a-881a-a739684f8bab" path="/var/lib/kubelet/pods/9155f175-88a3-462a-881a-a739684f8bab/volumes" Dec 27 08:40:52 crc kubenswrapper[4934]: I1227 08:40:52.063454 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wm2lb"] Dec 27 08:40:52 crc kubenswrapper[4934]: E1227 08:40:52.064504 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9155f175-88a3-462a-881a-a739684f8bab" containerName="extract-content" Dec 27 08:40:52 crc kubenswrapper[4934]: I1227 08:40:52.064520 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="9155f175-88a3-462a-881a-a739684f8bab" containerName="extract-content" Dec 27 08:40:52 crc kubenswrapper[4934]: E1227 08:40:52.064547 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9155f175-88a3-462a-881a-a739684f8bab" containerName="registry-server" Dec 27 08:40:52 crc kubenswrapper[4934]: I1227 08:40:52.064556 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="9155f175-88a3-462a-881a-a739684f8bab" containerName="registry-server" Dec 27 08:40:52 crc kubenswrapper[4934]: E1227 08:40:52.064595 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9155f175-88a3-462a-881a-a739684f8bab" containerName="extract-utilities" Dec 27 08:40:52 crc kubenswrapper[4934]: I1227 08:40:52.064603 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="9155f175-88a3-462a-881a-a739684f8bab" containerName="extract-utilities" Dec 27 08:40:52 crc kubenswrapper[4934]: I1227 08:40:52.064980 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="9155f175-88a3-462a-881a-a739684f8bab" containerName="registry-server" Dec 27 08:40:52 crc kubenswrapper[4934]: I1227 08:40:52.067036 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wm2lb" Dec 27 08:40:52 crc kubenswrapper[4934]: I1227 08:40:52.107037 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wm2lb"] Dec 27 08:40:52 crc kubenswrapper[4934]: I1227 08:40:52.171908 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/264cd7ab-5b9e-4c57-bfdf-d063110a95fe-utilities\") pod \"redhat-operators-wm2lb\" (UID: \"264cd7ab-5b9e-4c57-bfdf-d063110a95fe\") " pod="openshift-marketplace/redhat-operators-wm2lb" Dec 27 08:40:52 crc kubenswrapper[4934]: I1227 08:40:52.172024 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/264cd7ab-5b9e-4c57-bfdf-d063110a95fe-catalog-content\") pod \"redhat-operators-wm2lb\" (UID: \"264cd7ab-5b9e-4c57-bfdf-d063110a95fe\") " pod="openshift-marketplace/redhat-operators-wm2lb" Dec 27 08:40:52 crc kubenswrapper[4934]: I1227 08:40:52.172281 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvc4s\" (UniqueName: \"kubernetes.io/projected/264cd7ab-5b9e-4c57-bfdf-d063110a95fe-kube-api-access-lvc4s\") pod \"redhat-operators-wm2lb\" (UID: \"264cd7ab-5b9e-4c57-bfdf-d063110a95fe\") " pod="openshift-marketplace/redhat-operators-wm2lb" Dec 27 08:40:52 crc kubenswrapper[4934]: I1227 08:40:52.274020 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/264cd7ab-5b9e-4c57-bfdf-d063110a95fe-catalog-content\") pod \"redhat-operators-wm2lb\" (UID: \"264cd7ab-5b9e-4c57-bfdf-d063110a95fe\") " pod="openshift-marketplace/redhat-operators-wm2lb" Dec 27 08:40:52 crc kubenswrapper[4934]: I1227 08:40:52.274131 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvc4s\" (UniqueName: \"kubernetes.io/projected/264cd7ab-5b9e-4c57-bfdf-d063110a95fe-kube-api-access-lvc4s\") pod \"redhat-operators-wm2lb\" (UID: \"264cd7ab-5b9e-4c57-bfdf-d063110a95fe\") " pod="openshift-marketplace/redhat-operators-wm2lb" Dec 27 08:40:52 crc kubenswrapper[4934]: I1227 08:40:52.274235 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/264cd7ab-5b9e-4c57-bfdf-d063110a95fe-utilities\") pod \"redhat-operators-wm2lb\" (UID: \"264cd7ab-5b9e-4c57-bfdf-d063110a95fe\") " pod="openshift-marketplace/redhat-operators-wm2lb" Dec 27 08:40:52 crc kubenswrapper[4934]: I1227 08:40:52.274609 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/264cd7ab-5b9e-4c57-bfdf-d063110a95fe-catalog-content\") pod \"redhat-operators-wm2lb\" (UID: \"264cd7ab-5b9e-4c57-bfdf-d063110a95fe\") " pod="openshift-marketplace/redhat-operators-wm2lb" Dec 27 08:40:52 crc kubenswrapper[4934]: I1227 08:40:52.274662 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/264cd7ab-5b9e-4c57-bfdf-d063110a95fe-utilities\") pod \"redhat-operators-wm2lb\" (UID: \"264cd7ab-5b9e-4c57-bfdf-d063110a95fe\") " pod="openshift-marketplace/redhat-operators-wm2lb" Dec 27 08:40:52 crc kubenswrapper[4934]: I1227 08:40:52.299023 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvc4s\" (UniqueName: \"kubernetes.io/projected/264cd7ab-5b9e-4c57-bfdf-d063110a95fe-kube-api-access-lvc4s\") pod \"redhat-operators-wm2lb\" (UID: \"264cd7ab-5b9e-4c57-bfdf-d063110a95fe\") " pod="openshift-marketplace/redhat-operators-wm2lb" Dec 27 08:40:52 crc kubenswrapper[4934]: I1227 08:40:52.406755 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wm2lb" Dec 27 08:40:52 crc kubenswrapper[4934]: I1227 08:40:52.891665 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wm2lb"] Dec 27 08:40:52 crc kubenswrapper[4934]: W1227 08:40:52.896882 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod264cd7ab_5b9e_4c57_bfdf_d063110a95fe.slice/crio-e1690663a3fea703496e8cb2a813bcec0a41eba7f5ec84a9e10039d62205df5c WatchSource:0}: Error finding container e1690663a3fea703496e8cb2a813bcec0a41eba7f5ec84a9e10039d62205df5c: Status 404 returned error can't find the container with id e1690663a3fea703496e8cb2a813bcec0a41eba7f5ec84a9e10039d62205df5c Dec 27 08:40:53 crc kubenswrapper[4934]: I1227 08:40:53.100108 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wm2lb" event={"ID":"264cd7ab-5b9e-4c57-bfdf-d063110a95fe","Type":"ContainerStarted","Data":"e1690663a3fea703496e8cb2a813bcec0a41eba7f5ec84a9e10039d62205df5c"} Dec 27 08:40:54 crc kubenswrapper[4934]: I1227 08:40:54.118368 4934 generic.go:334] "Generic (PLEG): container finished" podID="264cd7ab-5b9e-4c57-bfdf-d063110a95fe" containerID="5a531f492eefb27d66c4313549e46e12a63cb081e93e78528fd7e846b6c89bff" exitCode=0 Dec 27 08:40:54 crc kubenswrapper[4934]: I1227 08:40:54.118474 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wm2lb" event={"ID":"264cd7ab-5b9e-4c57-bfdf-d063110a95fe","Type":"ContainerDied","Data":"5a531f492eefb27d66c4313549e46e12a63cb081e93e78528fd7e846b6c89bff"} Dec 27 08:40:55 crc kubenswrapper[4934]: I1227 08:40:55.158321 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wm2lb" event={"ID":"264cd7ab-5b9e-4c57-bfdf-d063110a95fe","Type":"ContainerStarted","Data":"e1d4cdc5f12c62a9d57866a1c4199e9b5f9c89152d092d84ad18c7dd2740bd0c"} Dec 27 08:40:59 crc kubenswrapper[4934]: I1227 08:40:59.220736 4934 generic.go:334] "Generic (PLEG): container finished" podID="264cd7ab-5b9e-4c57-bfdf-d063110a95fe" containerID="e1d4cdc5f12c62a9d57866a1c4199e9b5f9c89152d092d84ad18c7dd2740bd0c" exitCode=0 Dec 27 08:40:59 crc kubenswrapper[4934]: I1227 08:40:59.220817 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wm2lb" event={"ID":"264cd7ab-5b9e-4c57-bfdf-d063110a95fe","Type":"ContainerDied","Data":"e1d4cdc5f12c62a9d57866a1c4199e9b5f9c89152d092d84ad18c7dd2740bd0c"} Dec 27 08:41:01 crc kubenswrapper[4934]: I1227 08:41:01.249752 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wm2lb" event={"ID":"264cd7ab-5b9e-4c57-bfdf-d063110a95fe","Type":"ContainerStarted","Data":"718d2d50ebc52d8b5d7d2ff3020e791db2d0f5f6d17471be495a35b87fe3ba4b"} Dec 27 08:41:01 crc kubenswrapper[4934]: I1227 08:41:01.279137 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wm2lb" podStartSLOduration=3.524697712 podStartE2EDuration="9.279109799s" podCreationTimestamp="2025-12-27 08:40:52 +0000 UTC" firstStartedPulling="2025-12-27 08:40:54.120760264 +0000 UTC m=+3514.941200878" lastFinishedPulling="2025-12-27 08:40:59.875172331 +0000 UTC m=+3520.695612965" observedRunningTime="2025-12-27 08:41:01.274032005 +0000 UTC m=+3522.094472609" watchObservedRunningTime="2025-12-27 08:41:01.279109799 +0000 UTC m=+3522.099550413" Dec 27 08:41:02 crc kubenswrapper[4934]: I1227 08:41:02.407303 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wm2lb" Dec 27 08:41:02 crc kubenswrapper[4934]: I1227 08:41:02.407774 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wm2lb" Dec 27 08:41:03 crc kubenswrapper[4934]: I1227 08:41:03.471696 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wm2lb" podUID="264cd7ab-5b9e-4c57-bfdf-d063110a95fe" containerName="registry-server" probeResult="failure" output=< Dec 27 08:41:03 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 08:41:03 crc kubenswrapper[4934]: > Dec 27 08:41:12 crc kubenswrapper[4934]: I1227 08:41:12.471438 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wm2lb" Dec 27 08:41:12 crc kubenswrapper[4934]: I1227 08:41:12.543848 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wm2lb" Dec 27 08:41:13 crc kubenswrapper[4934]: I1227 08:41:13.380928 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wm2lb"] Dec 27 08:41:14 crc kubenswrapper[4934]: I1227 08:41:14.403587 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wm2lb" podUID="264cd7ab-5b9e-4c57-bfdf-d063110a95fe" containerName="registry-server" containerID="cri-o://718d2d50ebc52d8b5d7d2ff3020e791db2d0f5f6d17471be495a35b87fe3ba4b" gracePeriod=2 Dec 27 08:41:14 crc kubenswrapper[4934]: I1227 08:41:14.927001 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wm2lb" Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.116364 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/264cd7ab-5b9e-4c57-bfdf-d063110a95fe-catalog-content\") pod \"264cd7ab-5b9e-4c57-bfdf-d063110a95fe\" (UID: \"264cd7ab-5b9e-4c57-bfdf-d063110a95fe\") " Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.116604 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/264cd7ab-5b9e-4c57-bfdf-d063110a95fe-utilities\") pod \"264cd7ab-5b9e-4c57-bfdf-d063110a95fe\" (UID: \"264cd7ab-5b9e-4c57-bfdf-d063110a95fe\") " Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.116806 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvc4s\" (UniqueName: \"kubernetes.io/projected/264cd7ab-5b9e-4c57-bfdf-d063110a95fe-kube-api-access-lvc4s\") pod \"264cd7ab-5b9e-4c57-bfdf-d063110a95fe\" (UID: \"264cd7ab-5b9e-4c57-bfdf-d063110a95fe\") " Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.117352 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/264cd7ab-5b9e-4c57-bfdf-d063110a95fe-utilities" (OuterVolumeSpecName: "utilities") pod "264cd7ab-5b9e-4c57-bfdf-d063110a95fe" (UID: "264cd7ab-5b9e-4c57-bfdf-d063110a95fe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.117599 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/264cd7ab-5b9e-4c57-bfdf-d063110a95fe-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.125655 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/264cd7ab-5b9e-4c57-bfdf-d063110a95fe-kube-api-access-lvc4s" (OuterVolumeSpecName: "kube-api-access-lvc4s") pod "264cd7ab-5b9e-4c57-bfdf-d063110a95fe" (UID: "264cd7ab-5b9e-4c57-bfdf-d063110a95fe"). InnerVolumeSpecName "kube-api-access-lvc4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.220383 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvc4s\" (UniqueName: \"kubernetes.io/projected/264cd7ab-5b9e-4c57-bfdf-d063110a95fe-kube-api-access-lvc4s\") on node \"crc\" DevicePath \"\"" Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.232508 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/264cd7ab-5b9e-4c57-bfdf-d063110a95fe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "264cd7ab-5b9e-4c57-bfdf-d063110a95fe" (UID: "264cd7ab-5b9e-4c57-bfdf-d063110a95fe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.323512 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/264cd7ab-5b9e-4c57-bfdf-d063110a95fe-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.416155 4934 generic.go:334] "Generic (PLEG): container finished" podID="264cd7ab-5b9e-4c57-bfdf-d063110a95fe" containerID="718d2d50ebc52d8b5d7d2ff3020e791db2d0f5f6d17471be495a35b87fe3ba4b" exitCode=0 Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.416202 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wm2lb" event={"ID":"264cd7ab-5b9e-4c57-bfdf-d063110a95fe","Type":"ContainerDied","Data":"718d2d50ebc52d8b5d7d2ff3020e791db2d0f5f6d17471be495a35b87fe3ba4b"} Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.416234 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wm2lb" event={"ID":"264cd7ab-5b9e-4c57-bfdf-d063110a95fe","Type":"ContainerDied","Data":"e1690663a3fea703496e8cb2a813bcec0a41eba7f5ec84a9e10039d62205df5c"} Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.416251 4934 scope.go:117] "RemoveContainer" containerID="718d2d50ebc52d8b5d7d2ff3020e791db2d0f5f6d17471be495a35b87fe3ba4b" Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.416255 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wm2lb" Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.439885 4934 scope.go:117] "RemoveContainer" containerID="e1d4cdc5f12c62a9d57866a1c4199e9b5f9c89152d092d84ad18c7dd2740bd0c" Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.488160 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wm2lb"] Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.491253 4934 scope.go:117] "RemoveContainer" containerID="5a531f492eefb27d66c4313549e46e12a63cb081e93e78528fd7e846b6c89bff" Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.491996 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wm2lb"] Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.537142 4934 scope.go:117] "RemoveContainer" containerID="718d2d50ebc52d8b5d7d2ff3020e791db2d0f5f6d17471be495a35b87fe3ba4b" Dec 27 08:41:15 crc kubenswrapper[4934]: E1227 08:41:15.537764 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"718d2d50ebc52d8b5d7d2ff3020e791db2d0f5f6d17471be495a35b87fe3ba4b\": container with ID starting with 718d2d50ebc52d8b5d7d2ff3020e791db2d0f5f6d17471be495a35b87fe3ba4b not found: ID does not exist" containerID="718d2d50ebc52d8b5d7d2ff3020e791db2d0f5f6d17471be495a35b87fe3ba4b" Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.537854 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"718d2d50ebc52d8b5d7d2ff3020e791db2d0f5f6d17471be495a35b87fe3ba4b"} err="failed to get container status \"718d2d50ebc52d8b5d7d2ff3020e791db2d0f5f6d17471be495a35b87fe3ba4b\": rpc error: code = NotFound desc = could not find container \"718d2d50ebc52d8b5d7d2ff3020e791db2d0f5f6d17471be495a35b87fe3ba4b\": container with ID starting with 718d2d50ebc52d8b5d7d2ff3020e791db2d0f5f6d17471be495a35b87fe3ba4b not found: ID does not exist" Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.537903 4934 scope.go:117] "RemoveContainer" containerID="e1d4cdc5f12c62a9d57866a1c4199e9b5f9c89152d092d84ad18c7dd2740bd0c" Dec 27 08:41:15 crc kubenswrapper[4934]: E1227 08:41:15.538587 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1d4cdc5f12c62a9d57866a1c4199e9b5f9c89152d092d84ad18c7dd2740bd0c\": container with ID starting with e1d4cdc5f12c62a9d57866a1c4199e9b5f9c89152d092d84ad18c7dd2740bd0c not found: ID does not exist" containerID="e1d4cdc5f12c62a9d57866a1c4199e9b5f9c89152d092d84ad18c7dd2740bd0c" Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.538640 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1d4cdc5f12c62a9d57866a1c4199e9b5f9c89152d092d84ad18c7dd2740bd0c"} err="failed to get container status \"e1d4cdc5f12c62a9d57866a1c4199e9b5f9c89152d092d84ad18c7dd2740bd0c\": rpc error: code = NotFound desc = could not find container \"e1d4cdc5f12c62a9d57866a1c4199e9b5f9c89152d092d84ad18c7dd2740bd0c\": container with ID starting with e1d4cdc5f12c62a9d57866a1c4199e9b5f9c89152d092d84ad18c7dd2740bd0c not found: ID does not exist" Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.538668 4934 scope.go:117] "RemoveContainer" containerID="5a531f492eefb27d66c4313549e46e12a63cb081e93e78528fd7e846b6c89bff" Dec 27 08:41:15 crc kubenswrapper[4934]: E1227 08:41:15.539072 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a531f492eefb27d66c4313549e46e12a63cb081e93e78528fd7e846b6c89bff\": container with ID starting with 5a531f492eefb27d66c4313549e46e12a63cb081e93e78528fd7e846b6c89bff not found: ID does not exist" containerID="5a531f492eefb27d66c4313549e46e12a63cb081e93e78528fd7e846b6c89bff" Dec 27 08:41:15 crc kubenswrapper[4934]: I1227 08:41:15.539142 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a531f492eefb27d66c4313549e46e12a63cb081e93e78528fd7e846b6c89bff"} err="failed to get container status \"5a531f492eefb27d66c4313549e46e12a63cb081e93e78528fd7e846b6c89bff\": rpc error: code = NotFound desc = could not find container \"5a531f492eefb27d66c4313549e46e12a63cb081e93e78528fd7e846b6c89bff\": container with ID starting with 5a531f492eefb27d66c4313549e46e12a63cb081e93e78528fd7e846b6c89bff not found: ID does not exist" Dec 27 08:41:17 crc kubenswrapper[4934]: I1227 08:41:17.497608 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="264cd7ab-5b9e-4c57-bfdf-d063110a95fe" path="/var/lib/kubelet/pods/264cd7ab-5b9e-4c57-bfdf-d063110a95fe/volumes" Dec 27 08:42:15 crc kubenswrapper[4934]: I1227 08:42:15.329780 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:42:15 crc kubenswrapper[4934]: I1227 08:42:15.330370 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:42:45 crc kubenswrapper[4934]: I1227 08:42:45.330499 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:42:45 crc kubenswrapper[4934]: I1227 08:42:45.330879 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:43:15 crc kubenswrapper[4934]: I1227 08:43:15.330435 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:43:15 crc kubenswrapper[4934]: I1227 08:43:15.331032 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:43:15 crc kubenswrapper[4934]: I1227 08:43:15.331170 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 08:43:15 crc kubenswrapper[4934]: I1227 08:43:15.332500 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 08:43:15 crc kubenswrapper[4934]: I1227 08:43:15.332602 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" gracePeriod=600 Dec 27 08:43:15 crc kubenswrapper[4934]: E1227 08:43:15.459917 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:43:15 crc kubenswrapper[4934]: I1227 08:43:15.648122 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" exitCode=0 Dec 27 08:43:15 crc kubenswrapper[4934]: I1227 08:43:15.648590 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754"} Dec 27 08:43:15 crc kubenswrapper[4934]: I1227 08:43:15.648774 4934 scope.go:117] "RemoveContainer" containerID="2e3d7a11613b1624818a75601c039cc83fb1411488789f6d83381d669b30ac39" Dec 27 08:43:15 crc kubenswrapper[4934]: I1227 08:43:15.650660 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:43:15 crc kubenswrapper[4934]: E1227 08:43:15.652286 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:43:26 crc kubenswrapper[4934]: I1227 08:43:26.467769 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:43:26 crc kubenswrapper[4934]: E1227 08:43:26.468685 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:43:39 crc kubenswrapper[4934]: I1227 08:43:39.478636 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:43:39 crc kubenswrapper[4934]: E1227 08:43:39.479477 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:43:52 crc kubenswrapper[4934]: I1227 08:43:52.470523 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:43:52 crc kubenswrapper[4934]: E1227 08:43:52.472223 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:44:04 crc kubenswrapper[4934]: E1227 08:44:04.013870 4934 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.13:49724->38.102.83.13:43999: write tcp 38.102.83.13:49724->38.102.83.13:43999: write: broken pipe Dec 27 08:44:05 crc kubenswrapper[4934]: I1227 08:44:05.471035 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:44:05 crc kubenswrapper[4934]: E1227 08:44:05.471806 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:44:18 crc kubenswrapper[4934]: I1227 08:44:18.468393 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:44:18 crc kubenswrapper[4934]: E1227 08:44:18.469606 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:44:33 crc kubenswrapper[4934]: I1227 08:44:33.468533 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:44:33 crc kubenswrapper[4934]: E1227 08:44:33.469380 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:44:48 crc kubenswrapper[4934]: I1227 08:44:48.468738 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:44:48 crc kubenswrapper[4934]: E1227 08:44:48.470167 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:45:00 crc kubenswrapper[4934]: I1227 08:45:00.178031 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447085-mnn49"] Dec 27 08:45:00 crc kubenswrapper[4934]: E1227 08:45:00.179148 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="264cd7ab-5b9e-4c57-bfdf-d063110a95fe" containerName="extract-utilities" Dec 27 08:45:00 crc kubenswrapper[4934]: I1227 08:45:00.179165 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="264cd7ab-5b9e-4c57-bfdf-d063110a95fe" containerName="extract-utilities" Dec 27 08:45:00 crc kubenswrapper[4934]: E1227 08:45:00.179281 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="264cd7ab-5b9e-4c57-bfdf-d063110a95fe" containerName="extract-content" Dec 27 08:45:00 crc kubenswrapper[4934]: I1227 08:45:00.179294 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="264cd7ab-5b9e-4c57-bfdf-d063110a95fe" containerName="extract-content" Dec 27 08:45:00 crc kubenswrapper[4934]: E1227 08:45:00.179313 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="264cd7ab-5b9e-4c57-bfdf-d063110a95fe" containerName="registry-server" Dec 27 08:45:00 crc kubenswrapper[4934]: I1227 08:45:00.179320 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="264cd7ab-5b9e-4c57-bfdf-d063110a95fe" containerName="registry-server" Dec 27 08:45:00 crc kubenswrapper[4934]: I1227 08:45:00.179703 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="264cd7ab-5b9e-4c57-bfdf-d063110a95fe" containerName="registry-server" Dec 27 08:45:00 crc kubenswrapper[4934]: I1227 08:45:00.181703 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447085-mnn49" Dec 27 08:45:00 crc kubenswrapper[4934]: I1227 08:45:00.183922 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 27 08:45:00 crc kubenswrapper[4934]: I1227 08:45:00.184360 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 27 08:45:00 crc kubenswrapper[4934]: I1227 08:45:00.202584 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447085-mnn49"] Dec 27 08:45:00 crc kubenswrapper[4934]: I1227 08:45:00.262293 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/868f30f9-ef52-458b-a35c-193269bfba46-secret-volume\") pod \"collect-profiles-29447085-mnn49\" (UID: \"868f30f9-ef52-458b-a35c-193269bfba46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447085-mnn49" Dec 27 08:45:00 crc kubenswrapper[4934]: I1227 08:45:00.262527 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4rwl\" (UniqueName: \"kubernetes.io/projected/868f30f9-ef52-458b-a35c-193269bfba46-kube-api-access-s4rwl\") pod \"collect-profiles-29447085-mnn49\" (UID: \"868f30f9-ef52-458b-a35c-193269bfba46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447085-mnn49" Dec 27 08:45:00 crc kubenswrapper[4934]: I1227 08:45:00.262600 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/868f30f9-ef52-458b-a35c-193269bfba46-config-volume\") pod \"collect-profiles-29447085-mnn49\" (UID: \"868f30f9-ef52-458b-a35c-193269bfba46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447085-mnn49" Dec 27 08:45:00 crc kubenswrapper[4934]: I1227 08:45:00.365641 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/868f30f9-ef52-458b-a35c-193269bfba46-secret-volume\") pod \"collect-profiles-29447085-mnn49\" (UID: \"868f30f9-ef52-458b-a35c-193269bfba46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447085-mnn49" Dec 27 08:45:00 crc kubenswrapper[4934]: I1227 08:45:00.365729 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4rwl\" (UniqueName: \"kubernetes.io/projected/868f30f9-ef52-458b-a35c-193269bfba46-kube-api-access-s4rwl\") pod \"collect-profiles-29447085-mnn49\" (UID: \"868f30f9-ef52-458b-a35c-193269bfba46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447085-mnn49" Dec 27 08:45:00 crc kubenswrapper[4934]: I1227 08:45:00.365764 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/868f30f9-ef52-458b-a35c-193269bfba46-config-volume\") pod \"collect-profiles-29447085-mnn49\" (UID: \"868f30f9-ef52-458b-a35c-193269bfba46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447085-mnn49" Dec 27 08:45:00 crc kubenswrapper[4934]: I1227 08:45:00.366716 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/868f30f9-ef52-458b-a35c-193269bfba46-config-volume\") pod \"collect-profiles-29447085-mnn49\" (UID: \"868f30f9-ef52-458b-a35c-193269bfba46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447085-mnn49" Dec 27 08:45:00 crc kubenswrapper[4934]: I1227 08:45:00.378214 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/868f30f9-ef52-458b-a35c-193269bfba46-secret-volume\") pod \"collect-profiles-29447085-mnn49\" (UID: \"868f30f9-ef52-458b-a35c-193269bfba46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447085-mnn49" Dec 27 08:45:00 crc kubenswrapper[4934]: I1227 08:45:00.383904 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4rwl\" (UniqueName: \"kubernetes.io/projected/868f30f9-ef52-458b-a35c-193269bfba46-kube-api-access-s4rwl\") pod \"collect-profiles-29447085-mnn49\" (UID: \"868f30f9-ef52-458b-a35c-193269bfba46\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447085-mnn49" Dec 27 08:45:00 crc kubenswrapper[4934]: I1227 08:45:00.510471 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447085-mnn49" Dec 27 08:45:00 crc kubenswrapper[4934]: I1227 08:45:00.987920 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447085-mnn49"] Dec 27 08:45:01 crc kubenswrapper[4934]: I1227 08:45:01.162973 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447085-mnn49" event={"ID":"868f30f9-ef52-458b-a35c-193269bfba46","Type":"ContainerStarted","Data":"84719e0948d5e20b7a00e37c5e36459d858a23880b0cf500b664477fca5acc47"} Dec 27 08:45:02 crc kubenswrapper[4934]: I1227 08:45:02.180275 4934 generic.go:334] "Generic (PLEG): container finished" podID="868f30f9-ef52-458b-a35c-193269bfba46" containerID="82d5e1c7a3c2a693188e5e1754b115498b98c16a15a8af6a3bc44a1d55a7f3ef" exitCode=0 Dec 27 08:45:02 crc kubenswrapper[4934]: I1227 08:45:02.181351 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447085-mnn49" event={"ID":"868f30f9-ef52-458b-a35c-193269bfba46","Type":"ContainerDied","Data":"82d5e1c7a3c2a693188e5e1754b115498b98c16a15a8af6a3bc44a1d55a7f3ef"} Dec 27 08:45:02 crc kubenswrapper[4934]: I1227 08:45:02.467973 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:45:02 crc kubenswrapper[4934]: E1227 08:45:02.468636 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:45:03 crc kubenswrapper[4934]: I1227 08:45:03.625195 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447085-mnn49" Dec 27 08:45:03 crc kubenswrapper[4934]: I1227 08:45:03.649186 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4rwl\" (UniqueName: \"kubernetes.io/projected/868f30f9-ef52-458b-a35c-193269bfba46-kube-api-access-s4rwl\") pod \"868f30f9-ef52-458b-a35c-193269bfba46\" (UID: \"868f30f9-ef52-458b-a35c-193269bfba46\") " Dec 27 08:45:03 crc kubenswrapper[4934]: I1227 08:45:03.649388 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/868f30f9-ef52-458b-a35c-193269bfba46-secret-volume\") pod \"868f30f9-ef52-458b-a35c-193269bfba46\" (UID: \"868f30f9-ef52-458b-a35c-193269bfba46\") " Dec 27 08:45:03 crc kubenswrapper[4934]: I1227 08:45:03.649550 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/868f30f9-ef52-458b-a35c-193269bfba46-config-volume\") pod \"868f30f9-ef52-458b-a35c-193269bfba46\" (UID: \"868f30f9-ef52-458b-a35c-193269bfba46\") " Dec 27 08:45:03 crc kubenswrapper[4934]: I1227 08:45:03.651125 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/868f30f9-ef52-458b-a35c-193269bfba46-config-volume" (OuterVolumeSpecName: "config-volume") pod "868f30f9-ef52-458b-a35c-193269bfba46" (UID: "868f30f9-ef52-458b-a35c-193269bfba46"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 08:45:03 crc kubenswrapper[4934]: I1227 08:45:03.657452 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/868f30f9-ef52-458b-a35c-193269bfba46-kube-api-access-s4rwl" (OuterVolumeSpecName: "kube-api-access-s4rwl") pod "868f30f9-ef52-458b-a35c-193269bfba46" (UID: "868f30f9-ef52-458b-a35c-193269bfba46"). InnerVolumeSpecName "kube-api-access-s4rwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:45:03 crc kubenswrapper[4934]: I1227 08:45:03.661682 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/868f30f9-ef52-458b-a35c-193269bfba46-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "868f30f9-ef52-458b-a35c-193269bfba46" (UID: "868f30f9-ef52-458b-a35c-193269bfba46"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 08:45:03 crc kubenswrapper[4934]: I1227 08:45:03.753195 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4rwl\" (UniqueName: \"kubernetes.io/projected/868f30f9-ef52-458b-a35c-193269bfba46-kube-api-access-s4rwl\") on node \"crc\" DevicePath \"\"" Dec 27 08:45:03 crc kubenswrapper[4934]: I1227 08:45:03.753430 4934 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/868f30f9-ef52-458b-a35c-193269bfba46-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 27 08:45:03 crc kubenswrapper[4934]: I1227 08:45:03.753488 4934 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/868f30f9-ef52-458b-a35c-193269bfba46-config-volume\") on node \"crc\" DevicePath \"\"" Dec 27 08:45:04 crc kubenswrapper[4934]: I1227 08:45:04.204473 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447085-mnn49" event={"ID":"868f30f9-ef52-458b-a35c-193269bfba46","Type":"ContainerDied","Data":"84719e0948d5e20b7a00e37c5e36459d858a23880b0cf500b664477fca5acc47"} Dec 27 08:45:04 crc kubenswrapper[4934]: I1227 08:45:04.204767 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84719e0948d5e20b7a00e37c5e36459d858a23880b0cf500b664477fca5acc47" Dec 27 08:45:04 crc kubenswrapper[4934]: I1227 08:45:04.204605 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447085-mnn49" Dec 27 08:45:04 crc kubenswrapper[4934]: I1227 08:45:04.723204 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg"] Dec 27 08:45:04 crc kubenswrapper[4934]: I1227 08:45:04.735398 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447040-pw7gg"] Dec 27 08:45:05 crc kubenswrapper[4934]: I1227 08:45:05.492213 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5885f5e1-fac5-4980-8501-f882f612fa04" path="/var/lib/kubelet/pods/5885f5e1-fac5-4980-8501-f882f612fa04/volumes" Dec 27 08:45:14 crc kubenswrapper[4934]: I1227 08:45:14.468917 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:45:14 crc kubenswrapper[4934]: E1227 08:45:14.470578 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:45:29 crc kubenswrapper[4934]: I1227 08:45:29.206809 4934 scope.go:117] "RemoveContainer" containerID="e4c25105ac81a169026cb2bd288bf2ce3bb8524168682d6cdab900409c04d0f3" Dec 27 08:45:29 crc kubenswrapper[4934]: I1227 08:45:29.475552 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:45:29 crc kubenswrapper[4934]: E1227 08:45:29.475970 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:45:44 crc kubenswrapper[4934]: I1227 08:45:44.468596 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:45:44 crc kubenswrapper[4934]: E1227 08:45:44.469851 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:45:56 crc kubenswrapper[4934]: I1227 08:45:56.468041 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:45:56 crc kubenswrapper[4934]: E1227 08:45:56.469380 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:46:09 crc kubenswrapper[4934]: I1227 08:46:09.468483 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:46:09 crc kubenswrapper[4934]: E1227 08:46:09.469834 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:46:20 crc kubenswrapper[4934]: I1227 08:46:20.468555 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:46:20 crc kubenswrapper[4934]: E1227 08:46:20.469824 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:46:33 crc kubenswrapper[4934]: I1227 08:46:33.468202 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:46:33 crc kubenswrapper[4934]: E1227 08:46:33.468984 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:46:46 crc kubenswrapper[4934]: I1227 08:46:46.468606 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:46:46 crc kubenswrapper[4934]: E1227 08:46:46.469701 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:46:46 crc kubenswrapper[4934]: I1227 08:46:46.538799 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6jnzv"] Dec 27 08:46:46 crc kubenswrapper[4934]: E1227 08:46:46.539543 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="868f30f9-ef52-458b-a35c-193269bfba46" containerName="collect-profiles" Dec 27 08:46:46 crc kubenswrapper[4934]: I1227 08:46:46.539575 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="868f30f9-ef52-458b-a35c-193269bfba46" containerName="collect-profiles" Dec 27 08:46:46 crc kubenswrapper[4934]: I1227 08:46:46.539970 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="868f30f9-ef52-458b-a35c-193269bfba46" containerName="collect-profiles" Dec 27 08:46:46 crc kubenswrapper[4934]: I1227 08:46:46.544129 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6jnzv" Dec 27 08:46:46 crc kubenswrapper[4934]: I1227 08:46:46.551313 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6jnzv"] Dec 27 08:46:46 crc kubenswrapper[4934]: I1227 08:46:46.677071 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-279c4\" (UniqueName: \"kubernetes.io/projected/947f8acc-1c2c-4373-b184-7c895300ecaf-kube-api-access-279c4\") pod \"community-operators-6jnzv\" (UID: \"947f8acc-1c2c-4373-b184-7c895300ecaf\") " pod="openshift-marketplace/community-operators-6jnzv" Dec 27 08:46:46 crc kubenswrapper[4934]: I1227 08:46:46.677433 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/947f8acc-1c2c-4373-b184-7c895300ecaf-utilities\") pod \"community-operators-6jnzv\" (UID: \"947f8acc-1c2c-4373-b184-7c895300ecaf\") " pod="openshift-marketplace/community-operators-6jnzv" Dec 27 08:46:46 crc kubenswrapper[4934]: I1227 08:46:46.677505 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/947f8acc-1c2c-4373-b184-7c895300ecaf-catalog-content\") pod \"community-operators-6jnzv\" (UID: \"947f8acc-1c2c-4373-b184-7c895300ecaf\") " pod="openshift-marketplace/community-operators-6jnzv" Dec 27 08:46:46 crc kubenswrapper[4934]: I1227 08:46:46.780365 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/947f8acc-1c2c-4373-b184-7c895300ecaf-catalog-content\") pod \"community-operators-6jnzv\" (UID: \"947f8acc-1c2c-4373-b184-7c895300ecaf\") " pod="openshift-marketplace/community-operators-6jnzv" Dec 27 08:46:46 crc kubenswrapper[4934]: I1227 08:46:46.780724 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-279c4\" (UniqueName: \"kubernetes.io/projected/947f8acc-1c2c-4373-b184-7c895300ecaf-kube-api-access-279c4\") pod \"community-operators-6jnzv\" (UID: \"947f8acc-1c2c-4373-b184-7c895300ecaf\") " pod="openshift-marketplace/community-operators-6jnzv" Dec 27 08:46:46 crc kubenswrapper[4934]: I1227 08:46:46.780771 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/947f8acc-1c2c-4373-b184-7c895300ecaf-utilities\") pod \"community-operators-6jnzv\" (UID: \"947f8acc-1c2c-4373-b184-7c895300ecaf\") " pod="openshift-marketplace/community-operators-6jnzv" Dec 27 08:46:46 crc kubenswrapper[4934]: I1227 08:46:46.780912 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/947f8acc-1c2c-4373-b184-7c895300ecaf-catalog-content\") pod \"community-operators-6jnzv\" (UID: \"947f8acc-1c2c-4373-b184-7c895300ecaf\") " pod="openshift-marketplace/community-operators-6jnzv" Dec 27 08:46:46 crc kubenswrapper[4934]: I1227 08:46:46.781329 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/947f8acc-1c2c-4373-b184-7c895300ecaf-utilities\") pod \"community-operators-6jnzv\" (UID: \"947f8acc-1c2c-4373-b184-7c895300ecaf\") " pod="openshift-marketplace/community-operators-6jnzv" Dec 27 08:46:46 crc kubenswrapper[4934]: I1227 08:46:46.805108 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-279c4\" (UniqueName: \"kubernetes.io/projected/947f8acc-1c2c-4373-b184-7c895300ecaf-kube-api-access-279c4\") pod \"community-operators-6jnzv\" (UID: \"947f8acc-1c2c-4373-b184-7c895300ecaf\") " pod="openshift-marketplace/community-operators-6jnzv" Dec 27 08:46:46 crc kubenswrapper[4934]: I1227 08:46:46.874208 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6jnzv" Dec 27 08:46:47 crc kubenswrapper[4934]: I1227 08:46:47.660589 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6jnzv"] Dec 27 08:46:48 crc kubenswrapper[4934]: I1227 08:46:48.564399 4934 generic.go:334] "Generic (PLEG): container finished" podID="947f8acc-1c2c-4373-b184-7c895300ecaf" containerID="989bc1fdf21b937aa125df7f11f34b61e6ae8b7dac6d42265d558f5b8bbbc875" exitCode=0 Dec 27 08:46:48 crc kubenswrapper[4934]: I1227 08:46:48.564530 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jnzv" event={"ID":"947f8acc-1c2c-4373-b184-7c895300ecaf","Type":"ContainerDied","Data":"989bc1fdf21b937aa125df7f11f34b61e6ae8b7dac6d42265d558f5b8bbbc875"} Dec 27 08:46:48 crc kubenswrapper[4934]: I1227 08:46:48.564838 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jnzv" event={"ID":"947f8acc-1c2c-4373-b184-7c895300ecaf","Type":"ContainerStarted","Data":"eea014ce552b1ed07bb55f192ca7fee11e84e9e808af812e4f3fd06dcd1a1afd"} Dec 27 08:46:48 crc kubenswrapper[4934]: I1227 08:46:48.566821 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 27 08:46:49 crc kubenswrapper[4934]: I1227 08:46:49.578518 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jnzv" event={"ID":"947f8acc-1c2c-4373-b184-7c895300ecaf","Type":"ContainerStarted","Data":"dae8b1afd40a48c1f007d0bc0b73143f5c4b434fd4a5780489d6403f1fc11386"} Dec 27 08:46:50 crc kubenswrapper[4934]: I1227 08:46:50.594999 4934 generic.go:334] "Generic (PLEG): container finished" podID="947f8acc-1c2c-4373-b184-7c895300ecaf" containerID="dae8b1afd40a48c1f007d0bc0b73143f5c4b434fd4a5780489d6403f1fc11386" exitCode=0 Dec 27 08:46:50 crc kubenswrapper[4934]: I1227 08:46:50.595157 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jnzv" event={"ID":"947f8acc-1c2c-4373-b184-7c895300ecaf","Type":"ContainerDied","Data":"dae8b1afd40a48c1f007d0bc0b73143f5c4b434fd4a5780489d6403f1fc11386"} Dec 27 08:46:51 crc kubenswrapper[4934]: I1227 08:46:51.609941 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jnzv" event={"ID":"947f8acc-1c2c-4373-b184-7c895300ecaf","Type":"ContainerStarted","Data":"bef7800ca2585fe4cb4972dfa6a433e748c5cda482ac776d7a8e0fb0315cd2c1"} Dec 27 08:46:51 crc kubenswrapper[4934]: I1227 08:46:51.639871 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6jnzv" podStartSLOduration=3.086531473 podStartE2EDuration="5.639854805s" podCreationTimestamp="2025-12-27 08:46:46 +0000 UTC" firstStartedPulling="2025-12-27 08:46:48.566593995 +0000 UTC m=+3869.387034589" lastFinishedPulling="2025-12-27 08:46:51.119917327 +0000 UTC m=+3871.940357921" observedRunningTime="2025-12-27 08:46:51.634323558 +0000 UTC m=+3872.454764152" watchObservedRunningTime="2025-12-27 08:46:51.639854805 +0000 UTC m=+3872.460295399" Dec 27 08:46:56 crc kubenswrapper[4934]: I1227 08:46:56.874888 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6jnzv" Dec 27 08:46:56 crc kubenswrapper[4934]: I1227 08:46:56.875454 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6jnzv" Dec 27 08:46:56 crc kubenswrapper[4934]: I1227 08:46:56.942494 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6jnzv" Dec 27 08:46:57 crc kubenswrapper[4934]: I1227 08:46:57.743615 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6jnzv" Dec 27 08:46:57 crc kubenswrapper[4934]: I1227 08:46:57.799598 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6jnzv"] Dec 27 08:46:59 crc kubenswrapper[4934]: I1227 08:46:59.706713 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6jnzv" podUID="947f8acc-1c2c-4373-b184-7c895300ecaf" containerName="registry-server" containerID="cri-o://bef7800ca2585fe4cb4972dfa6a433e748c5cda482ac776d7a8e0fb0315cd2c1" gracePeriod=2 Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.275165 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6jnzv" Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.319431 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/947f8acc-1c2c-4373-b184-7c895300ecaf-utilities\") pod \"947f8acc-1c2c-4373-b184-7c895300ecaf\" (UID: \"947f8acc-1c2c-4373-b184-7c895300ecaf\") " Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.319760 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279c4\" (UniqueName: \"kubernetes.io/projected/947f8acc-1c2c-4373-b184-7c895300ecaf-kube-api-access-279c4\") pod \"947f8acc-1c2c-4373-b184-7c895300ecaf\" (UID: \"947f8acc-1c2c-4373-b184-7c895300ecaf\") " Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.319828 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/947f8acc-1c2c-4373-b184-7c895300ecaf-catalog-content\") pod \"947f8acc-1c2c-4373-b184-7c895300ecaf\" (UID: \"947f8acc-1c2c-4373-b184-7c895300ecaf\") " Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.320845 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/947f8acc-1c2c-4373-b184-7c895300ecaf-utilities" (OuterVolumeSpecName: "utilities") pod "947f8acc-1c2c-4373-b184-7c895300ecaf" (UID: "947f8acc-1c2c-4373-b184-7c895300ecaf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.330423 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/947f8acc-1c2c-4373-b184-7c895300ecaf-kube-api-access-279c4" (OuterVolumeSpecName: "kube-api-access-279c4") pod "947f8acc-1c2c-4373-b184-7c895300ecaf" (UID: "947f8acc-1c2c-4373-b184-7c895300ecaf"). InnerVolumeSpecName "kube-api-access-279c4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.407103 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/947f8acc-1c2c-4373-b184-7c895300ecaf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "947f8acc-1c2c-4373-b184-7c895300ecaf" (UID: "947f8acc-1c2c-4373-b184-7c895300ecaf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.422654 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279c4\" (UniqueName: \"kubernetes.io/projected/947f8acc-1c2c-4373-b184-7c895300ecaf-kube-api-access-279c4\") on node \"crc\" DevicePath \"\"" Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.422689 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/947f8acc-1c2c-4373-b184-7c895300ecaf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.422702 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/947f8acc-1c2c-4373-b184-7c895300ecaf-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.718995 4934 generic.go:334] "Generic (PLEG): container finished" podID="947f8acc-1c2c-4373-b184-7c895300ecaf" containerID="bef7800ca2585fe4cb4972dfa6a433e748c5cda482ac776d7a8e0fb0315cd2c1" exitCode=0 Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.719052 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6jnzv" Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.719061 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jnzv" event={"ID":"947f8acc-1c2c-4373-b184-7c895300ecaf","Type":"ContainerDied","Data":"bef7800ca2585fe4cb4972dfa6a433e748c5cda482ac776d7a8e0fb0315cd2c1"} Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.719135 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6jnzv" event={"ID":"947f8acc-1c2c-4373-b184-7c895300ecaf","Type":"ContainerDied","Data":"eea014ce552b1ed07bb55f192ca7fee11e84e9e808af812e4f3fd06dcd1a1afd"} Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.719162 4934 scope.go:117] "RemoveContainer" containerID="bef7800ca2585fe4cb4972dfa6a433e748c5cda482ac776d7a8e0fb0315cd2c1" Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.747231 4934 scope.go:117] "RemoveContainer" containerID="dae8b1afd40a48c1f007d0bc0b73143f5c4b434fd4a5780489d6403f1fc11386" Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.756142 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6jnzv"] Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.766932 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6jnzv"] Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.798780 4934 scope.go:117] "RemoveContainer" containerID="989bc1fdf21b937aa125df7f11f34b61e6ae8b7dac6d42265d558f5b8bbbc875" Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.831459 4934 scope.go:117] "RemoveContainer" containerID="bef7800ca2585fe4cb4972dfa6a433e748c5cda482ac776d7a8e0fb0315cd2c1" Dec 27 08:47:00 crc kubenswrapper[4934]: E1227 08:47:00.832421 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bef7800ca2585fe4cb4972dfa6a433e748c5cda482ac776d7a8e0fb0315cd2c1\": container with ID starting with bef7800ca2585fe4cb4972dfa6a433e748c5cda482ac776d7a8e0fb0315cd2c1 not found: ID does not exist" containerID="bef7800ca2585fe4cb4972dfa6a433e748c5cda482ac776d7a8e0fb0315cd2c1" Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.832467 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bef7800ca2585fe4cb4972dfa6a433e748c5cda482ac776d7a8e0fb0315cd2c1"} err="failed to get container status \"bef7800ca2585fe4cb4972dfa6a433e748c5cda482ac776d7a8e0fb0315cd2c1\": rpc error: code = NotFound desc = could not find container \"bef7800ca2585fe4cb4972dfa6a433e748c5cda482ac776d7a8e0fb0315cd2c1\": container with ID starting with bef7800ca2585fe4cb4972dfa6a433e748c5cda482ac776d7a8e0fb0315cd2c1 not found: ID does not exist" Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.832496 4934 scope.go:117] "RemoveContainer" containerID="dae8b1afd40a48c1f007d0bc0b73143f5c4b434fd4a5780489d6403f1fc11386" Dec 27 08:47:00 crc kubenswrapper[4934]: E1227 08:47:00.832886 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dae8b1afd40a48c1f007d0bc0b73143f5c4b434fd4a5780489d6403f1fc11386\": container with ID starting with dae8b1afd40a48c1f007d0bc0b73143f5c4b434fd4a5780489d6403f1fc11386 not found: ID does not exist" containerID="dae8b1afd40a48c1f007d0bc0b73143f5c4b434fd4a5780489d6403f1fc11386" Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.832928 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dae8b1afd40a48c1f007d0bc0b73143f5c4b434fd4a5780489d6403f1fc11386"} err="failed to get container status \"dae8b1afd40a48c1f007d0bc0b73143f5c4b434fd4a5780489d6403f1fc11386\": rpc error: code = NotFound desc = could not find container \"dae8b1afd40a48c1f007d0bc0b73143f5c4b434fd4a5780489d6403f1fc11386\": container with ID starting with dae8b1afd40a48c1f007d0bc0b73143f5c4b434fd4a5780489d6403f1fc11386 not found: ID does not exist" Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.832958 4934 scope.go:117] "RemoveContainer" containerID="989bc1fdf21b937aa125df7f11f34b61e6ae8b7dac6d42265d558f5b8bbbc875" Dec 27 08:47:00 crc kubenswrapper[4934]: E1227 08:47:00.833429 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"989bc1fdf21b937aa125df7f11f34b61e6ae8b7dac6d42265d558f5b8bbbc875\": container with ID starting with 989bc1fdf21b937aa125df7f11f34b61e6ae8b7dac6d42265d558f5b8bbbc875 not found: ID does not exist" containerID="989bc1fdf21b937aa125df7f11f34b61e6ae8b7dac6d42265d558f5b8bbbc875" Dec 27 08:47:00 crc kubenswrapper[4934]: I1227 08:47:00.833454 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"989bc1fdf21b937aa125df7f11f34b61e6ae8b7dac6d42265d558f5b8bbbc875"} err="failed to get container status \"989bc1fdf21b937aa125df7f11f34b61e6ae8b7dac6d42265d558f5b8bbbc875\": rpc error: code = NotFound desc = could not find container \"989bc1fdf21b937aa125df7f11f34b61e6ae8b7dac6d42265d558f5b8bbbc875\": container with ID starting with 989bc1fdf21b937aa125df7f11f34b61e6ae8b7dac6d42265d558f5b8bbbc875 not found: ID does not exist" Dec 27 08:47:01 crc kubenswrapper[4934]: I1227 08:47:01.468141 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:47:01 crc kubenswrapper[4934]: E1227 08:47:01.468617 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:47:01 crc kubenswrapper[4934]: I1227 08:47:01.488825 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="947f8acc-1c2c-4373-b184-7c895300ecaf" path="/var/lib/kubelet/pods/947f8acc-1c2c-4373-b184-7c895300ecaf/volumes" Dec 27 08:47:12 crc kubenswrapper[4934]: I1227 08:47:12.467389 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:47:12 crc kubenswrapper[4934]: E1227 08:47:12.468150 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:47:26 crc kubenswrapper[4934]: I1227 08:47:26.466892 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:47:26 crc kubenswrapper[4934]: E1227 08:47:26.468462 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:47:40 crc kubenswrapper[4934]: I1227 08:47:40.468102 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:47:40 crc kubenswrapper[4934]: E1227 08:47:40.469028 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:47:55 crc kubenswrapper[4934]: I1227 08:47:55.468756 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:47:55 crc kubenswrapper[4934]: E1227 08:47:55.469585 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:48:06 crc kubenswrapper[4934]: I1227 08:48:06.468019 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:48:06 crc kubenswrapper[4934]: E1227 08:48:06.469069 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:48:19 crc kubenswrapper[4934]: I1227 08:48:19.478619 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:48:20 crc kubenswrapper[4934]: I1227 08:48:20.173409 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"c468952548ae16cc22785c126991e9b5b77b35573edaa125b08c1c9d98f746ac"} Dec 27 08:50:22 crc kubenswrapper[4934]: I1227 08:50:22.832202 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-znppn"] Dec 27 08:50:22 crc kubenswrapper[4934]: E1227 08:50:22.833066 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="947f8acc-1c2c-4373-b184-7c895300ecaf" containerName="registry-server" Dec 27 08:50:22 crc kubenswrapper[4934]: I1227 08:50:22.833081 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="947f8acc-1c2c-4373-b184-7c895300ecaf" containerName="registry-server" Dec 27 08:50:22 crc kubenswrapper[4934]: E1227 08:50:22.833139 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="947f8acc-1c2c-4373-b184-7c895300ecaf" containerName="extract-utilities" Dec 27 08:50:22 crc kubenswrapper[4934]: I1227 08:50:22.833145 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="947f8acc-1c2c-4373-b184-7c895300ecaf" containerName="extract-utilities" Dec 27 08:50:22 crc kubenswrapper[4934]: E1227 08:50:22.833167 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="947f8acc-1c2c-4373-b184-7c895300ecaf" containerName="extract-content" Dec 27 08:50:22 crc kubenswrapper[4934]: I1227 08:50:22.833174 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="947f8acc-1c2c-4373-b184-7c895300ecaf" containerName="extract-content" Dec 27 08:50:22 crc kubenswrapper[4934]: I1227 08:50:22.833406 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="947f8acc-1c2c-4373-b184-7c895300ecaf" containerName="registry-server" Dec 27 08:50:22 crc kubenswrapper[4934]: I1227 08:50:22.835335 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-znppn" Dec 27 08:50:22 crc kubenswrapper[4934]: I1227 08:50:22.851357 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-znppn"] Dec 27 08:50:22 crc kubenswrapper[4934]: I1227 08:50:22.961719 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abe05313-1f50-4eff-80ff-d801a1f97d75-utilities\") pod \"certified-operators-znppn\" (UID: \"abe05313-1f50-4eff-80ff-d801a1f97d75\") " pod="openshift-marketplace/certified-operators-znppn" Dec 27 08:50:22 crc kubenswrapper[4934]: I1227 08:50:22.961800 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abe05313-1f50-4eff-80ff-d801a1f97d75-catalog-content\") pod \"certified-operators-znppn\" (UID: \"abe05313-1f50-4eff-80ff-d801a1f97d75\") " pod="openshift-marketplace/certified-operators-znppn" Dec 27 08:50:22 crc kubenswrapper[4934]: I1227 08:50:22.961850 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pckrk\" (UniqueName: \"kubernetes.io/projected/abe05313-1f50-4eff-80ff-d801a1f97d75-kube-api-access-pckrk\") pod \"certified-operators-znppn\" (UID: \"abe05313-1f50-4eff-80ff-d801a1f97d75\") " pod="openshift-marketplace/certified-operators-znppn" Dec 27 08:50:23 crc kubenswrapper[4934]: I1227 08:50:23.064556 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abe05313-1f50-4eff-80ff-d801a1f97d75-utilities\") pod \"certified-operators-znppn\" (UID: \"abe05313-1f50-4eff-80ff-d801a1f97d75\") " pod="openshift-marketplace/certified-operators-znppn" Dec 27 08:50:23 crc kubenswrapper[4934]: I1227 08:50:23.064682 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abe05313-1f50-4eff-80ff-d801a1f97d75-catalog-content\") pod \"certified-operators-znppn\" (UID: \"abe05313-1f50-4eff-80ff-d801a1f97d75\") " pod="openshift-marketplace/certified-operators-znppn" Dec 27 08:50:23 crc kubenswrapper[4934]: I1227 08:50:23.064754 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pckrk\" (UniqueName: \"kubernetes.io/projected/abe05313-1f50-4eff-80ff-d801a1f97d75-kube-api-access-pckrk\") pod \"certified-operators-znppn\" (UID: \"abe05313-1f50-4eff-80ff-d801a1f97d75\") " pod="openshift-marketplace/certified-operators-znppn" Dec 27 08:50:23 crc kubenswrapper[4934]: I1227 08:50:23.065387 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abe05313-1f50-4eff-80ff-d801a1f97d75-utilities\") pod \"certified-operators-znppn\" (UID: \"abe05313-1f50-4eff-80ff-d801a1f97d75\") " pod="openshift-marketplace/certified-operators-znppn" Dec 27 08:50:23 crc kubenswrapper[4934]: I1227 08:50:23.065646 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abe05313-1f50-4eff-80ff-d801a1f97d75-catalog-content\") pod \"certified-operators-znppn\" (UID: \"abe05313-1f50-4eff-80ff-d801a1f97d75\") " pod="openshift-marketplace/certified-operators-znppn" Dec 27 08:50:23 crc kubenswrapper[4934]: I1227 08:50:23.096009 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pckrk\" (UniqueName: \"kubernetes.io/projected/abe05313-1f50-4eff-80ff-d801a1f97d75-kube-api-access-pckrk\") pod \"certified-operators-znppn\" (UID: \"abe05313-1f50-4eff-80ff-d801a1f97d75\") " pod="openshift-marketplace/certified-operators-znppn" Dec 27 08:50:23 crc kubenswrapper[4934]: I1227 08:50:23.155241 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-znppn" Dec 27 08:50:23 crc kubenswrapper[4934]: I1227 08:50:23.687070 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-znppn"] Dec 27 08:50:24 crc kubenswrapper[4934]: I1227 08:50:24.703489 4934 generic.go:334] "Generic (PLEG): container finished" podID="abe05313-1f50-4eff-80ff-d801a1f97d75" containerID="c3541da0f9f99b5797bef6e9e572ab23770bc074b8a370084e19d664c6673c3c" exitCode=0 Dec 27 08:50:24 crc kubenswrapper[4934]: I1227 08:50:24.703577 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znppn" event={"ID":"abe05313-1f50-4eff-80ff-d801a1f97d75","Type":"ContainerDied","Data":"c3541da0f9f99b5797bef6e9e572ab23770bc074b8a370084e19d664c6673c3c"} Dec 27 08:50:24 crc kubenswrapper[4934]: I1227 08:50:24.704109 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znppn" event={"ID":"abe05313-1f50-4eff-80ff-d801a1f97d75","Type":"ContainerStarted","Data":"83d9f4056d13e9984f9f8dd0314ec7a20e31bd98aaadaa53d0efc3b725294c14"} Dec 27 08:50:25 crc kubenswrapper[4934]: I1227 08:50:25.725764 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znppn" event={"ID":"abe05313-1f50-4eff-80ff-d801a1f97d75","Type":"ContainerStarted","Data":"2a4a98029b03a83e88df3a3c3f05ac1ef94e475633622cbe4a79120c28869abc"} Dec 27 08:50:27 crc kubenswrapper[4934]: I1227 08:50:27.757108 4934 generic.go:334] "Generic (PLEG): container finished" podID="abe05313-1f50-4eff-80ff-d801a1f97d75" containerID="2a4a98029b03a83e88df3a3c3f05ac1ef94e475633622cbe4a79120c28869abc" exitCode=0 Dec 27 08:50:27 crc kubenswrapper[4934]: I1227 08:50:27.757764 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znppn" event={"ID":"abe05313-1f50-4eff-80ff-d801a1f97d75","Type":"ContainerDied","Data":"2a4a98029b03a83e88df3a3c3f05ac1ef94e475633622cbe4a79120c28869abc"} Dec 27 08:50:28 crc kubenswrapper[4934]: I1227 08:50:28.781578 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znppn" event={"ID":"abe05313-1f50-4eff-80ff-d801a1f97d75","Type":"ContainerStarted","Data":"f84a39c3c396ffccd3f0203527d1e35f63de5cd13b76f8b22c3f2077ee8f5bb1"} Dec 27 08:50:28 crc kubenswrapper[4934]: I1227 08:50:28.813543 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-znppn" podStartSLOduration=3.115850228 podStartE2EDuration="6.81352188s" podCreationTimestamp="2025-12-27 08:50:22 +0000 UTC" firstStartedPulling="2025-12-27 08:50:24.705203253 +0000 UTC m=+4085.525643847" lastFinishedPulling="2025-12-27 08:50:28.402874905 +0000 UTC m=+4089.223315499" observedRunningTime="2025-12-27 08:50:28.804043386 +0000 UTC m=+4089.624483990" watchObservedRunningTime="2025-12-27 08:50:28.81352188 +0000 UTC m=+4089.633962474" Dec 27 08:50:33 crc kubenswrapper[4934]: I1227 08:50:33.156572 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-znppn" Dec 27 08:50:33 crc kubenswrapper[4934]: I1227 08:50:33.157040 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-znppn" Dec 27 08:50:33 crc kubenswrapper[4934]: I1227 08:50:33.210049 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-znppn" Dec 27 08:50:33 crc kubenswrapper[4934]: I1227 08:50:33.905302 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-znppn" Dec 27 08:50:33 crc kubenswrapper[4934]: I1227 08:50:33.982195 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-znppn"] Dec 27 08:50:35 crc kubenswrapper[4934]: I1227 08:50:35.861119 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-znppn" podUID="abe05313-1f50-4eff-80ff-d801a1f97d75" containerName="registry-server" containerID="cri-o://f84a39c3c396ffccd3f0203527d1e35f63de5cd13b76f8b22c3f2077ee8f5bb1" gracePeriod=2 Dec 27 08:50:36 crc kubenswrapper[4934]: I1227 08:50:36.468142 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-znppn" Dec 27 08:50:36 crc kubenswrapper[4934]: I1227 08:50:36.630619 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abe05313-1f50-4eff-80ff-d801a1f97d75-utilities\") pod \"abe05313-1f50-4eff-80ff-d801a1f97d75\" (UID: \"abe05313-1f50-4eff-80ff-d801a1f97d75\") " Dec 27 08:50:36 crc kubenswrapper[4934]: I1227 08:50:36.630693 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pckrk\" (UniqueName: \"kubernetes.io/projected/abe05313-1f50-4eff-80ff-d801a1f97d75-kube-api-access-pckrk\") pod \"abe05313-1f50-4eff-80ff-d801a1f97d75\" (UID: \"abe05313-1f50-4eff-80ff-d801a1f97d75\") " Dec 27 08:50:36 crc kubenswrapper[4934]: I1227 08:50:36.630983 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abe05313-1f50-4eff-80ff-d801a1f97d75-catalog-content\") pod \"abe05313-1f50-4eff-80ff-d801a1f97d75\" (UID: \"abe05313-1f50-4eff-80ff-d801a1f97d75\") " Dec 27 08:50:36 crc kubenswrapper[4934]: I1227 08:50:36.631586 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abe05313-1f50-4eff-80ff-d801a1f97d75-utilities" (OuterVolumeSpecName: "utilities") pod "abe05313-1f50-4eff-80ff-d801a1f97d75" (UID: "abe05313-1f50-4eff-80ff-d801a1f97d75"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:50:36 crc kubenswrapper[4934]: I1227 08:50:36.631998 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abe05313-1f50-4eff-80ff-d801a1f97d75-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 08:50:36 crc kubenswrapper[4934]: I1227 08:50:36.689593 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/abe05313-1f50-4eff-80ff-d801a1f97d75-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "abe05313-1f50-4eff-80ff-d801a1f97d75" (UID: "abe05313-1f50-4eff-80ff-d801a1f97d75"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:50:36 crc kubenswrapper[4934]: I1227 08:50:36.734003 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abe05313-1f50-4eff-80ff-d801a1f97d75-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 08:50:37 crc kubenswrapper[4934]: I1227 08:50:37.005295 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abe05313-1f50-4eff-80ff-d801a1f97d75-kube-api-access-pckrk" (OuterVolumeSpecName: "kube-api-access-pckrk") pod "abe05313-1f50-4eff-80ff-d801a1f97d75" (UID: "abe05313-1f50-4eff-80ff-d801a1f97d75"). InnerVolumeSpecName "kube-api-access-pckrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:50:37 crc kubenswrapper[4934]: I1227 08:50:37.008901 4934 generic.go:334] "Generic (PLEG): container finished" podID="abe05313-1f50-4eff-80ff-d801a1f97d75" containerID="f84a39c3c396ffccd3f0203527d1e35f63de5cd13b76f8b22c3f2077ee8f5bb1" exitCode=0 Dec 27 08:50:37 crc kubenswrapper[4934]: I1227 08:50:37.008952 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znppn" event={"ID":"abe05313-1f50-4eff-80ff-d801a1f97d75","Type":"ContainerDied","Data":"f84a39c3c396ffccd3f0203527d1e35f63de5cd13b76f8b22c3f2077ee8f5bb1"} Dec 27 08:50:37 crc kubenswrapper[4934]: I1227 08:50:37.008984 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-znppn" event={"ID":"abe05313-1f50-4eff-80ff-d801a1f97d75","Type":"ContainerDied","Data":"83d9f4056d13e9984f9f8dd0314ec7a20e31bd98aaadaa53d0efc3b725294c14"} Dec 27 08:50:37 crc kubenswrapper[4934]: I1227 08:50:37.009008 4934 scope.go:117] "RemoveContainer" containerID="f84a39c3c396ffccd3f0203527d1e35f63de5cd13b76f8b22c3f2077ee8f5bb1" Dec 27 08:50:37 crc kubenswrapper[4934]: I1227 08:50:37.009202 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-znppn" Dec 27 08:50:37 crc kubenswrapper[4934]: I1227 08:50:37.039660 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pckrk\" (UniqueName: \"kubernetes.io/projected/abe05313-1f50-4eff-80ff-d801a1f97d75-kube-api-access-pckrk\") on node \"crc\" DevicePath \"\"" Dec 27 08:50:37 crc kubenswrapper[4934]: I1227 08:50:37.080993 4934 scope.go:117] "RemoveContainer" containerID="2a4a98029b03a83e88df3a3c3f05ac1ef94e475633622cbe4a79120c28869abc" Dec 27 08:50:37 crc kubenswrapper[4934]: I1227 08:50:37.092536 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-znppn"] Dec 27 08:50:37 crc kubenswrapper[4934]: I1227 08:50:37.107698 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-znppn"] Dec 27 08:50:37 crc kubenswrapper[4934]: I1227 08:50:37.117567 4934 scope.go:117] "RemoveContainer" containerID="c3541da0f9f99b5797bef6e9e572ab23770bc074b8a370084e19d664c6673c3c" Dec 27 08:50:37 crc kubenswrapper[4934]: I1227 08:50:37.162871 4934 scope.go:117] "RemoveContainer" containerID="f84a39c3c396ffccd3f0203527d1e35f63de5cd13b76f8b22c3f2077ee8f5bb1" Dec 27 08:50:37 crc kubenswrapper[4934]: E1227 08:50:37.163433 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f84a39c3c396ffccd3f0203527d1e35f63de5cd13b76f8b22c3f2077ee8f5bb1\": container with ID starting with f84a39c3c396ffccd3f0203527d1e35f63de5cd13b76f8b22c3f2077ee8f5bb1 not found: ID does not exist" containerID="f84a39c3c396ffccd3f0203527d1e35f63de5cd13b76f8b22c3f2077ee8f5bb1" Dec 27 08:50:37 crc kubenswrapper[4934]: I1227 08:50:37.163485 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f84a39c3c396ffccd3f0203527d1e35f63de5cd13b76f8b22c3f2077ee8f5bb1"} err="failed to get container status \"f84a39c3c396ffccd3f0203527d1e35f63de5cd13b76f8b22c3f2077ee8f5bb1\": rpc error: code = NotFound desc = could not find container \"f84a39c3c396ffccd3f0203527d1e35f63de5cd13b76f8b22c3f2077ee8f5bb1\": container with ID starting with f84a39c3c396ffccd3f0203527d1e35f63de5cd13b76f8b22c3f2077ee8f5bb1 not found: ID does not exist" Dec 27 08:50:37 crc kubenswrapper[4934]: I1227 08:50:37.163533 4934 scope.go:117] "RemoveContainer" containerID="2a4a98029b03a83e88df3a3c3f05ac1ef94e475633622cbe4a79120c28869abc" Dec 27 08:50:37 crc kubenswrapper[4934]: E1227 08:50:37.164161 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a4a98029b03a83e88df3a3c3f05ac1ef94e475633622cbe4a79120c28869abc\": container with ID starting with 2a4a98029b03a83e88df3a3c3f05ac1ef94e475633622cbe4a79120c28869abc not found: ID does not exist" containerID="2a4a98029b03a83e88df3a3c3f05ac1ef94e475633622cbe4a79120c28869abc" Dec 27 08:50:37 crc kubenswrapper[4934]: I1227 08:50:37.164204 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a4a98029b03a83e88df3a3c3f05ac1ef94e475633622cbe4a79120c28869abc"} err="failed to get container status \"2a4a98029b03a83e88df3a3c3f05ac1ef94e475633622cbe4a79120c28869abc\": rpc error: code = NotFound desc = could not find container \"2a4a98029b03a83e88df3a3c3f05ac1ef94e475633622cbe4a79120c28869abc\": container with ID starting with 2a4a98029b03a83e88df3a3c3f05ac1ef94e475633622cbe4a79120c28869abc not found: ID does not exist" Dec 27 08:50:37 crc kubenswrapper[4934]: I1227 08:50:37.164232 4934 scope.go:117] "RemoveContainer" containerID="c3541da0f9f99b5797bef6e9e572ab23770bc074b8a370084e19d664c6673c3c" Dec 27 08:50:37 crc kubenswrapper[4934]: E1227 08:50:37.164617 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3541da0f9f99b5797bef6e9e572ab23770bc074b8a370084e19d664c6673c3c\": container with ID starting with c3541da0f9f99b5797bef6e9e572ab23770bc074b8a370084e19d664c6673c3c not found: ID does not exist" containerID="c3541da0f9f99b5797bef6e9e572ab23770bc074b8a370084e19d664c6673c3c" Dec 27 08:50:37 crc kubenswrapper[4934]: I1227 08:50:37.164642 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3541da0f9f99b5797bef6e9e572ab23770bc074b8a370084e19d664c6673c3c"} err="failed to get container status \"c3541da0f9f99b5797bef6e9e572ab23770bc074b8a370084e19d664c6673c3c\": rpc error: code = NotFound desc = could not find container \"c3541da0f9f99b5797bef6e9e572ab23770bc074b8a370084e19d664c6673c3c\": container with ID starting with c3541da0f9f99b5797bef6e9e572ab23770bc074b8a370084e19d664c6673c3c not found: ID does not exist" Dec 27 08:50:37 crc kubenswrapper[4934]: I1227 08:50:37.487332 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abe05313-1f50-4eff-80ff-d801a1f97d75" path="/var/lib/kubelet/pods/abe05313-1f50-4eff-80ff-d801a1f97d75/volumes" Dec 27 08:50:45 crc kubenswrapper[4934]: I1227 08:50:45.330196 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:50:45 crc kubenswrapper[4934]: I1227 08:50:45.330778 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:51:15 crc kubenswrapper[4934]: I1227 08:51:15.330299 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:51:15 crc kubenswrapper[4934]: I1227 08:51:15.331413 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:51:45 crc kubenswrapper[4934]: I1227 08:51:45.329679 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:51:45 crc kubenswrapper[4934]: I1227 08:51:45.330955 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:51:45 crc kubenswrapper[4934]: I1227 08:51:45.331062 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 08:51:45 crc kubenswrapper[4934]: I1227 08:51:45.332270 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c468952548ae16cc22785c126991e9b5b77b35573edaa125b08c1c9d98f746ac"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 08:51:45 crc kubenswrapper[4934]: I1227 08:51:45.332356 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://c468952548ae16cc22785c126991e9b5b77b35573edaa125b08c1c9d98f746ac" gracePeriod=600 Dec 27 08:51:45 crc kubenswrapper[4934]: I1227 08:51:45.922962 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="c468952548ae16cc22785c126991e9b5b77b35573edaa125b08c1c9d98f746ac" exitCode=0 Dec 27 08:51:45 crc kubenswrapper[4934]: I1227 08:51:45.923048 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"c468952548ae16cc22785c126991e9b5b77b35573edaa125b08c1c9d98f746ac"} Dec 27 08:51:45 crc kubenswrapper[4934]: I1227 08:51:45.923286 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84"} Dec 27 08:51:45 crc kubenswrapper[4934]: I1227 08:51:45.923311 4934 scope.go:117] "RemoveContainer" containerID="5073fea416918a102c2ed0515291ff775c494442759c852e1ae42b4637bf0754" Dec 27 08:51:51 crc kubenswrapper[4934]: I1227 08:51:51.682788 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kq7vm"] Dec 27 08:51:51 crc kubenswrapper[4934]: E1227 08:51:51.683788 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abe05313-1f50-4eff-80ff-d801a1f97d75" containerName="extract-content" Dec 27 08:51:51 crc kubenswrapper[4934]: I1227 08:51:51.683800 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="abe05313-1f50-4eff-80ff-d801a1f97d75" containerName="extract-content" Dec 27 08:51:51 crc kubenswrapper[4934]: E1227 08:51:51.683822 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abe05313-1f50-4eff-80ff-d801a1f97d75" containerName="extract-utilities" Dec 27 08:51:51 crc kubenswrapper[4934]: I1227 08:51:51.683828 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="abe05313-1f50-4eff-80ff-d801a1f97d75" containerName="extract-utilities" Dec 27 08:51:51 crc kubenswrapper[4934]: E1227 08:51:51.683845 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abe05313-1f50-4eff-80ff-d801a1f97d75" containerName="registry-server" Dec 27 08:51:51 crc kubenswrapper[4934]: I1227 08:51:51.683851 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="abe05313-1f50-4eff-80ff-d801a1f97d75" containerName="registry-server" Dec 27 08:51:51 crc kubenswrapper[4934]: I1227 08:51:51.684078 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="abe05313-1f50-4eff-80ff-d801a1f97d75" containerName="registry-server" Dec 27 08:51:51 crc kubenswrapper[4934]: I1227 08:51:51.685730 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kq7vm" Dec 27 08:51:51 crc kubenswrapper[4934]: I1227 08:51:51.700127 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kq7vm"] Dec 27 08:51:51 crc kubenswrapper[4934]: I1227 08:51:51.792579 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/452266aa-3ebe-438a-adbe-722797eb887b-utilities\") pod \"redhat-operators-kq7vm\" (UID: \"452266aa-3ebe-438a-adbe-722797eb887b\") " pod="openshift-marketplace/redhat-operators-kq7vm" Dec 27 08:51:51 crc kubenswrapper[4934]: I1227 08:51:51.792695 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/452266aa-3ebe-438a-adbe-722797eb887b-catalog-content\") pod \"redhat-operators-kq7vm\" (UID: \"452266aa-3ebe-438a-adbe-722797eb887b\") " pod="openshift-marketplace/redhat-operators-kq7vm" Dec 27 08:51:51 crc kubenswrapper[4934]: I1227 08:51:51.792724 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpd96\" (UniqueName: \"kubernetes.io/projected/452266aa-3ebe-438a-adbe-722797eb887b-kube-api-access-rpd96\") pod \"redhat-operators-kq7vm\" (UID: \"452266aa-3ebe-438a-adbe-722797eb887b\") " pod="openshift-marketplace/redhat-operators-kq7vm" Dec 27 08:51:51 crc kubenswrapper[4934]: I1227 08:51:51.895143 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpd96\" (UniqueName: \"kubernetes.io/projected/452266aa-3ebe-438a-adbe-722797eb887b-kube-api-access-rpd96\") pod \"redhat-operators-kq7vm\" (UID: \"452266aa-3ebe-438a-adbe-722797eb887b\") " pod="openshift-marketplace/redhat-operators-kq7vm" Dec 27 08:51:51 crc kubenswrapper[4934]: I1227 08:51:51.895730 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/452266aa-3ebe-438a-adbe-722797eb887b-utilities\") pod \"redhat-operators-kq7vm\" (UID: \"452266aa-3ebe-438a-adbe-722797eb887b\") " pod="openshift-marketplace/redhat-operators-kq7vm" Dec 27 08:51:51 crc kubenswrapper[4934]: I1227 08:51:51.896235 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/452266aa-3ebe-438a-adbe-722797eb887b-utilities\") pod \"redhat-operators-kq7vm\" (UID: \"452266aa-3ebe-438a-adbe-722797eb887b\") " pod="openshift-marketplace/redhat-operators-kq7vm" Dec 27 08:51:51 crc kubenswrapper[4934]: I1227 08:51:51.896402 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/452266aa-3ebe-438a-adbe-722797eb887b-catalog-content\") pod \"redhat-operators-kq7vm\" (UID: \"452266aa-3ebe-438a-adbe-722797eb887b\") " pod="openshift-marketplace/redhat-operators-kq7vm" Dec 27 08:51:51 crc kubenswrapper[4934]: I1227 08:51:51.896671 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/452266aa-3ebe-438a-adbe-722797eb887b-catalog-content\") pod \"redhat-operators-kq7vm\" (UID: \"452266aa-3ebe-438a-adbe-722797eb887b\") " pod="openshift-marketplace/redhat-operators-kq7vm" Dec 27 08:51:51 crc kubenswrapper[4934]: I1227 08:51:51.918173 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpd96\" (UniqueName: \"kubernetes.io/projected/452266aa-3ebe-438a-adbe-722797eb887b-kube-api-access-rpd96\") pod \"redhat-operators-kq7vm\" (UID: \"452266aa-3ebe-438a-adbe-722797eb887b\") " pod="openshift-marketplace/redhat-operators-kq7vm" Dec 27 08:51:52 crc kubenswrapper[4934]: I1227 08:51:52.008701 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kq7vm" Dec 27 08:51:52 crc kubenswrapper[4934]: I1227 08:51:52.561422 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kq7vm"] Dec 27 08:51:53 crc kubenswrapper[4934]: I1227 08:51:53.010482 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq7vm" event={"ID":"452266aa-3ebe-438a-adbe-722797eb887b","Type":"ContainerStarted","Data":"088f3ec8bcc22928d94e306b4b69e0188288591b876113ec70600225259ee689"} Dec 27 08:51:53 crc kubenswrapper[4934]: I1227 08:51:53.010836 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq7vm" event={"ID":"452266aa-3ebe-438a-adbe-722797eb887b","Type":"ContainerStarted","Data":"07b773af596ee01a601c601b4b53d2d1099d7d1d077a53f7dc7ed2359cd6a580"} Dec 27 08:51:54 crc kubenswrapper[4934]: I1227 08:51:54.023205 4934 generic.go:334] "Generic (PLEG): container finished" podID="452266aa-3ebe-438a-adbe-722797eb887b" containerID="088f3ec8bcc22928d94e306b4b69e0188288591b876113ec70600225259ee689" exitCode=0 Dec 27 08:51:54 crc kubenswrapper[4934]: I1227 08:51:54.023288 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq7vm" event={"ID":"452266aa-3ebe-438a-adbe-722797eb887b","Type":"ContainerDied","Data":"088f3ec8bcc22928d94e306b4b69e0188288591b876113ec70600225259ee689"} Dec 27 08:51:54 crc kubenswrapper[4934]: I1227 08:51:54.026454 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 27 08:51:55 crc kubenswrapper[4934]: I1227 08:51:55.051440 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq7vm" event={"ID":"452266aa-3ebe-438a-adbe-722797eb887b","Type":"ContainerStarted","Data":"ec81cdbb31aabe9dfe0f73e021dfb187a64aebb340749ca6edb579927ad234fe"} Dec 27 08:51:59 crc kubenswrapper[4934]: I1227 08:51:59.144515 4934 generic.go:334] "Generic (PLEG): container finished" podID="452266aa-3ebe-438a-adbe-722797eb887b" containerID="ec81cdbb31aabe9dfe0f73e021dfb187a64aebb340749ca6edb579927ad234fe" exitCode=0 Dec 27 08:51:59 crc kubenswrapper[4934]: I1227 08:51:59.145056 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq7vm" event={"ID":"452266aa-3ebe-438a-adbe-722797eb887b","Type":"ContainerDied","Data":"ec81cdbb31aabe9dfe0f73e021dfb187a64aebb340749ca6edb579927ad234fe"} Dec 27 08:52:00 crc kubenswrapper[4934]: I1227 08:52:00.162589 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq7vm" event={"ID":"452266aa-3ebe-438a-adbe-722797eb887b","Type":"ContainerStarted","Data":"1cfffe830fd1b70beaa177ab5c07d2e3afd1356f8d6a75fd4068b87fe0556e86"} Dec 27 08:52:00 crc kubenswrapper[4934]: I1227 08:52:00.200525 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kq7vm" podStartSLOduration=3.649366826 podStartE2EDuration="9.200500893s" podCreationTimestamp="2025-12-27 08:51:51 +0000 UTC" firstStartedPulling="2025-12-27 08:51:54.026038051 +0000 UTC m=+4174.846478685" lastFinishedPulling="2025-12-27 08:51:59.577172118 +0000 UTC m=+4180.397612752" observedRunningTime="2025-12-27 08:52:00.186687881 +0000 UTC m=+4181.007128495" watchObservedRunningTime="2025-12-27 08:52:00.200500893 +0000 UTC m=+4181.020941487" Dec 27 08:52:02 crc kubenswrapper[4934]: I1227 08:52:02.009845 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kq7vm" Dec 27 08:52:02 crc kubenswrapper[4934]: I1227 08:52:02.010244 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kq7vm" Dec 27 08:52:03 crc kubenswrapper[4934]: I1227 08:52:03.062493 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kq7vm" podUID="452266aa-3ebe-438a-adbe-722797eb887b" containerName="registry-server" probeResult="failure" output=< Dec 27 08:52:03 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 08:52:03 crc kubenswrapper[4934]: > Dec 27 08:52:12 crc kubenswrapper[4934]: I1227 08:52:12.070398 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kq7vm" Dec 27 08:52:12 crc kubenswrapper[4934]: I1227 08:52:12.149524 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kq7vm" Dec 27 08:52:12 crc kubenswrapper[4934]: I1227 08:52:12.320054 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kq7vm"] Dec 27 08:52:13 crc kubenswrapper[4934]: I1227 08:52:13.329226 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kq7vm" podUID="452266aa-3ebe-438a-adbe-722797eb887b" containerName="registry-server" containerID="cri-o://1cfffe830fd1b70beaa177ab5c07d2e3afd1356f8d6a75fd4068b87fe0556e86" gracePeriod=2 Dec 27 08:52:13 crc kubenswrapper[4934]: I1227 08:52:13.946681 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kq7vm" Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.076161 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/452266aa-3ebe-438a-adbe-722797eb887b-catalog-content\") pod \"452266aa-3ebe-438a-adbe-722797eb887b\" (UID: \"452266aa-3ebe-438a-adbe-722797eb887b\") " Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.076209 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/452266aa-3ebe-438a-adbe-722797eb887b-utilities\") pod \"452266aa-3ebe-438a-adbe-722797eb887b\" (UID: \"452266aa-3ebe-438a-adbe-722797eb887b\") " Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.076443 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpd96\" (UniqueName: \"kubernetes.io/projected/452266aa-3ebe-438a-adbe-722797eb887b-kube-api-access-rpd96\") pod \"452266aa-3ebe-438a-adbe-722797eb887b\" (UID: \"452266aa-3ebe-438a-adbe-722797eb887b\") " Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.077566 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/452266aa-3ebe-438a-adbe-722797eb887b-utilities" (OuterVolumeSpecName: "utilities") pod "452266aa-3ebe-438a-adbe-722797eb887b" (UID: "452266aa-3ebe-438a-adbe-722797eb887b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.081919 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/452266aa-3ebe-438a-adbe-722797eb887b-kube-api-access-rpd96" (OuterVolumeSpecName: "kube-api-access-rpd96") pod "452266aa-3ebe-438a-adbe-722797eb887b" (UID: "452266aa-3ebe-438a-adbe-722797eb887b"). InnerVolumeSpecName "kube-api-access-rpd96". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.178923 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpd96\" (UniqueName: \"kubernetes.io/projected/452266aa-3ebe-438a-adbe-722797eb887b-kube-api-access-rpd96\") on node \"crc\" DevicePath \"\"" Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.178958 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/452266aa-3ebe-438a-adbe-722797eb887b-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.186432 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/452266aa-3ebe-438a-adbe-722797eb887b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "452266aa-3ebe-438a-adbe-722797eb887b" (UID: "452266aa-3ebe-438a-adbe-722797eb887b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.280844 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/452266aa-3ebe-438a-adbe-722797eb887b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.343511 4934 generic.go:334] "Generic (PLEG): container finished" podID="452266aa-3ebe-438a-adbe-722797eb887b" containerID="1cfffe830fd1b70beaa177ab5c07d2e3afd1356f8d6a75fd4068b87fe0556e86" exitCode=0 Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.343569 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq7vm" event={"ID":"452266aa-3ebe-438a-adbe-722797eb887b","Type":"ContainerDied","Data":"1cfffe830fd1b70beaa177ab5c07d2e3afd1356f8d6a75fd4068b87fe0556e86"} Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.343581 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kq7vm" Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.343613 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kq7vm" event={"ID":"452266aa-3ebe-438a-adbe-722797eb887b","Type":"ContainerDied","Data":"07b773af596ee01a601c601b4b53d2d1099d7d1d077a53f7dc7ed2359cd6a580"} Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.343641 4934 scope.go:117] "RemoveContainer" containerID="1cfffe830fd1b70beaa177ab5c07d2e3afd1356f8d6a75fd4068b87fe0556e86" Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.375978 4934 scope.go:117] "RemoveContainer" containerID="ec81cdbb31aabe9dfe0f73e021dfb187a64aebb340749ca6edb579927ad234fe" Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.381031 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kq7vm"] Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.393341 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kq7vm"] Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.401219 4934 scope.go:117] "RemoveContainer" containerID="088f3ec8bcc22928d94e306b4b69e0188288591b876113ec70600225259ee689" Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.466002 4934 scope.go:117] "RemoveContainer" containerID="1cfffe830fd1b70beaa177ab5c07d2e3afd1356f8d6a75fd4068b87fe0556e86" Dec 27 08:52:14 crc kubenswrapper[4934]: E1227 08:52:14.466523 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cfffe830fd1b70beaa177ab5c07d2e3afd1356f8d6a75fd4068b87fe0556e86\": container with ID starting with 1cfffe830fd1b70beaa177ab5c07d2e3afd1356f8d6a75fd4068b87fe0556e86 not found: ID does not exist" containerID="1cfffe830fd1b70beaa177ab5c07d2e3afd1356f8d6a75fd4068b87fe0556e86" Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.466628 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cfffe830fd1b70beaa177ab5c07d2e3afd1356f8d6a75fd4068b87fe0556e86"} err="failed to get container status \"1cfffe830fd1b70beaa177ab5c07d2e3afd1356f8d6a75fd4068b87fe0556e86\": rpc error: code = NotFound desc = could not find container \"1cfffe830fd1b70beaa177ab5c07d2e3afd1356f8d6a75fd4068b87fe0556e86\": container with ID starting with 1cfffe830fd1b70beaa177ab5c07d2e3afd1356f8d6a75fd4068b87fe0556e86 not found: ID does not exist" Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.466686 4934 scope.go:117] "RemoveContainer" containerID="ec81cdbb31aabe9dfe0f73e021dfb187a64aebb340749ca6edb579927ad234fe" Dec 27 08:52:14 crc kubenswrapper[4934]: E1227 08:52:14.467262 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec81cdbb31aabe9dfe0f73e021dfb187a64aebb340749ca6edb579927ad234fe\": container with ID starting with ec81cdbb31aabe9dfe0f73e021dfb187a64aebb340749ca6edb579927ad234fe not found: ID does not exist" containerID="ec81cdbb31aabe9dfe0f73e021dfb187a64aebb340749ca6edb579927ad234fe" Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.467367 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec81cdbb31aabe9dfe0f73e021dfb187a64aebb340749ca6edb579927ad234fe"} err="failed to get container status \"ec81cdbb31aabe9dfe0f73e021dfb187a64aebb340749ca6edb579927ad234fe\": rpc error: code = NotFound desc = could not find container \"ec81cdbb31aabe9dfe0f73e021dfb187a64aebb340749ca6edb579927ad234fe\": container with ID starting with ec81cdbb31aabe9dfe0f73e021dfb187a64aebb340749ca6edb579927ad234fe not found: ID does not exist" Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.467445 4934 scope.go:117] "RemoveContainer" containerID="088f3ec8bcc22928d94e306b4b69e0188288591b876113ec70600225259ee689" Dec 27 08:52:14 crc kubenswrapper[4934]: E1227 08:52:14.467852 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"088f3ec8bcc22928d94e306b4b69e0188288591b876113ec70600225259ee689\": container with ID starting with 088f3ec8bcc22928d94e306b4b69e0188288591b876113ec70600225259ee689 not found: ID does not exist" containerID="088f3ec8bcc22928d94e306b4b69e0188288591b876113ec70600225259ee689" Dec 27 08:52:14 crc kubenswrapper[4934]: I1227 08:52:14.467891 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"088f3ec8bcc22928d94e306b4b69e0188288591b876113ec70600225259ee689"} err="failed to get container status \"088f3ec8bcc22928d94e306b4b69e0188288591b876113ec70600225259ee689\": rpc error: code = NotFound desc = could not find container \"088f3ec8bcc22928d94e306b4b69e0188288591b876113ec70600225259ee689\": container with ID starting with 088f3ec8bcc22928d94e306b4b69e0188288591b876113ec70600225259ee689 not found: ID does not exist" Dec 27 08:52:15 crc kubenswrapper[4934]: I1227 08:52:15.479852 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="452266aa-3ebe-438a-adbe-722797eb887b" path="/var/lib/kubelet/pods/452266aa-3ebe-438a-adbe-722797eb887b/volumes" Dec 27 08:52:29 crc kubenswrapper[4934]: I1227 08:52:29.449183 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5m94h"] Dec 27 08:52:29 crc kubenswrapper[4934]: E1227 08:52:29.450543 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="452266aa-3ebe-438a-adbe-722797eb887b" containerName="registry-server" Dec 27 08:52:29 crc kubenswrapper[4934]: I1227 08:52:29.450562 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="452266aa-3ebe-438a-adbe-722797eb887b" containerName="registry-server" Dec 27 08:52:29 crc kubenswrapper[4934]: E1227 08:52:29.450586 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="452266aa-3ebe-438a-adbe-722797eb887b" containerName="extract-utilities" Dec 27 08:52:29 crc kubenswrapper[4934]: I1227 08:52:29.450595 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="452266aa-3ebe-438a-adbe-722797eb887b" containerName="extract-utilities" Dec 27 08:52:29 crc kubenswrapper[4934]: E1227 08:52:29.450645 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="452266aa-3ebe-438a-adbe-722797eb887b" containerName="extract-content" Dec 27 08:52:29 crc kubenswrapper[4934]: I1227 08:52:29.450653 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="452266aa-3ebe-438a-adbe-722797eb887b" containerName="extract-content" Dec 27 08:52:29 crc kubenswrapper[4934]: I1227 08:52:29.451108 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="452266aa-3ebe-438a-adbe-722797eb887b" containerName="registry-server" Dec 27 08:52:29 crc kubenswrapper[4934]: I1227 08:52:29.453490 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5m94h" Dec 27 08:52:29 crc kubenswrapper[4934]: I1227 08:52:29.463880 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5m94h"] Dec 27 08:52:29 crc kubenswrapper[4934]: I1227 08:52:29.571034 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b35f241-312f-4245-94f1-baf253d603bf-utilities\") pod \"redhat-marketplace-5m94h\" (UID: \"8b35f241-312f-4245-94f1-baf253d603bf\") " pod="openshift-marketplace/redhat-marketplace-5m94h" Dec 27 08:52:29 crc kubenswrapper[4934]: I1227 08:52:29.571219 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8njr\" (UniqueName: \"kubernetes.io/projected/8b35f241-312f-4245-94f1-baf253d603bf-kube-api-access-b8njr\") pod \"redhat-marketplace-5m94h\" (UID: \"8b35f241-312f-4245-94f1-baf253d603bf\") " pod="openshift-marketplace/redhat-marketplace-5m94h" Dec 27 08:52:29 crc kubenswrapper[4934]: I1227 08:52:29.571357 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b35f241-312f-4245-94f1-baf253d603bf-catalog-content\") pod \"redhat-marketplace-5m94h\" (UID: \"8b35f241-312f-4245-94f1-baf253d603bf\") " pod="openshift-marketplace/redhat-marketplace-5m94h" Dec 27 08:52:29 crc kubenswrapper[4934]: I1227 08:52:29.673858 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b35f241-312f-4245-94f1-baf253d603bf-utilities\") pod \"redhat-marketplace-5m94h\" (UID: \"8b35f241-312f-4245-94f1-baf253d603bf\") " pod="openshift-marketplace/redhat-marketplace-5m94h" Dec 27 08:52:29 crc kubenswrapper[4934]: I1227 08:52:29.673935 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8njr\" (UniqueName: \"kubernetes.io/projected/8b35f241-312f-4245-94f1-baf253d603bf-kube-api-access-b8njr\") pod \"redhat-marketplace-5m94h\" (UID: \"8b35f241-312f-4245-94f1-baf253d603bf\") " pod="openshift-marketplace/redhat-marketplace-5m94h" Dec 27 08:52:29 crc kubenswrapper[4934]: I1227 08:52:29.673987 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b35f241-312f-4245-94f1-baf253d603bf-catalog-content\") pod \"redhat-marketplace-5m94h\" (UID: \"8b35f241-312f-4245-94f1-baf253d603bf\") " pod="openshift-marketplace/redhat-marketplace-5m94h" Dec 27 08:52:29 crc kubenswrapper[4934]: I1227 08:52:29.674520 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b35f241-312f-4245-94f1-baf253d603bf-utilities\") pod \"redhat-marketplace-5m94h\" (UID: \"8b35f241-312f-4245-94f1-baf253d603bf\") " pod="openshift-marketplace/redhat-marketplace-5m94h" Dec 27 08:52:29 crc kubenswrapper[4934]: I1227 08:52:29.674581 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b35f241-312f-4245-94f1-baf253d603bf-catalog-content\") pod \"redhat-marketplace-5m94h\" (UID: \"8b35f241-312f-4245-94f1-baf253d603bf\") " pod="openshift-marketplace/redhat-marketplace-5m94h" Dec 27 08:52:29 crc kubenswrapper[4934]: I1227 08:52:29.692893 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8njr\" (UniqueName: \"kubernetes.io/projected/8b35f241-312f-4245-94f1-baf253d603bf-kube-api-access-b8njr\") pod \"redhat-marketplace-5m94h\" (UID: \"8b35f241-312f-4245-94f1-baf253d603bf\") " pod="openshift-marketplace/redhat-marketplace-5m94h" Dec 27 08:52:29 crc kubenswrapper[4934]: I1227 08:52:29.786312 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5m94h" Dec 27 08:52:30 crc kubenswrapper[4934]: I1227 08:52:30.288566 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5m94h"] Dec 27 08:52:31 crc kubenswrapper[4934]: I1227 08:52:31.548778 4934 generic.go:334] "Generic (PLEG): container finished" podID="8b35f241-312f-4245-94f1-baf253d603bf" containerID="10bd634712e3b7a6815ad769496b68ee138254d5d4f936f4b5f44669630a4a48" exitCode=0 Dec 27 08:52:31 crc kubenswrapper[4934]: I1227 08:52:31.548817 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5m94h" event={"ID":"8b35f241-312f-4245-94f1-baf253d603bf","Type":"ContainerDied","Data":"10bd634712e3b7a6815ad769496b68ee138254d5d4f936f4b5f44669630a4a48"} Dec 27 08:52:31 crc kubenswrapper[4934]: I1227 08:52:31.549056 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5m94h" event={"ID":"8b35f241-312f-4245-94f1-baf253d603bf","Type":"ContainerStarted","Data":"c9ead30f4eda8c996534cc2472d3d1952b696af419d9e2671b47c7bdb7abc084"} Dec 27 08:52:33 crc kubenswrapper[4934]: I1227 08:52:33.570689 4934 generic.go:334] "Generic (PLEG): container finished" podID="8b35f241-312f-4245-94f1-baf253d603bf" containerID="96e05ae1a9be16e7dd9fae3b841864e14dfeb187c0653e01490f67e3b32e9e2d" exitCode=0 Dec 27 08:52:33 crc kubenswrapper[4934]: I1227 08:52:33.570764 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5m94h" event={"ID":"8b35f241-312f-4245-94f1-baf253d603bf","Type":"ContainerDied","Data":"96e05ae1a9be16e7dd9fae3b841864e14dfeb187c0653e01490f67e3b32e9e2d"} Dec 27 08:52:34 crc kubenswrapper[4934]: I1227 08:52:34.586347 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5m94h" event={"ID":"8b35f241-312f-4245-94f1-baf253d603bf","Type":"ContainerStarted","Data":"ba8c3f5e5dec3fb1e4b72defdf9f150348c79064c27a61eb5ff8514aea6b71d0"} Dec 27 08:52:34 crc kubenswrapper[4934]: I1227 08:52:34.610421 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5m94h" podStartSLOduration=3.223914057 podStartE2EDuration="5.610396933s" podCreationTimestamp="2025-12-27 08:52:29 +0000 UTC" firstStartedPulling="2025-12-27 08:52:31.550577075 +0000 UTC m=+4212.371017669" lastFinishedPulling="2025-12-27 08:52:33.937059941 +0000 UTC m=+4214.757500545" observedRunningTime="2025-12-27 08:52:34.604551748 +0000 UTC m=+4215.424992362" watchObservedRunningTime="2025-12-27 08:52:34.610396933 +0000 UTC m=+4215.430837527" Dec 27 08:52:38 crc kubenswrapper[4934]: E1227 08:52:38.674143 4934 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.13:55098->38.102.83.13:43999: write tcp 38.102.83.13:55098->38.102.83.13:43999: write: broken pipe Dec 27 08:52:39 crc kubenswrapper[4934]: I1227 08:52:39.786624 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5m94h" Dec 27 08:52:39 crc kubenswrapper[4934]: I1227 08:52:39.787050 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5m94h" Dec 27 08:52:39 crc kubenswrapper[4934]: I1227 08:52:39.842161 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5m94h" Dec 27 08:52:40 crc kubenswrapper[4934]: I1227 08:52:40.180319 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5m94h" Dec 27 08:52:40 crc kubenswrapper[4934]: I1227 08:52:40.233861 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5m94h"] Dec 27 08:52:42 crc kubenswrapper[4934]: I1227 08:52:42.106522 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5m94h" podUID="8b35f241-312f-4245-94f1-baf253d603bf" containerName="registry-server" containerID="cri-o://ba8c3f5e5dec3fb1e4b72defdf9f150348c79064c27a61eb5ff8514aea6b71d0" gracePeriod=2 Dec 27 08:52:43 crc kubenswrapper[4934]: I1227 08:52:43.126498 4934 generic.go:334] "Generic (PLEG): container finished" podID="8b35f241-312f-4245-94f1-baf253d603bf" containerID="ba8c3f5e5dec3fb1e4b72defdf9f150348c79064c27a61eb5ff8514aea6b71d0" exitCode=0 Dec 27 08:52:43 crc kubenswrapper[4934]: I1227 08:52:43.126573 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5m94h" event={"ID":"8b35f241-312f-4245-94f1-baf253d603bf","Type":"ContainerDied","Data":"ba8c3f5e5dec3fb1e4b72defdf9f150348c79064c27a61eb5ff8514aea6b71d0"} Dec 27 08:52:43 crc kubenswrapper[4934]: I1227 08:52:43.126862 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5m94h" event={"ID":"8b35f241-312f-4245-94f1-baf253d603bf","Type":"ContainerDied","Data":"c9ead30f4eda8c996534cc2472d3d1952b696af419d9e2671b47c7bdb7abc084"} Dec 27 08:52:43 crc kubenswrapper[4934]: I1227 08:52:43.126882 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9ead30f4eda8c996534cc2472d3d1952b696af419d9e2671b47c7bdb7abc084" Dec 27 08:52:43 crc kubenswrapper[4934]: I1227 08:52:43.183101 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5m94h" Dec 27 08:52:43 crc kubenswrapper[4934]: I1227 08:52:43.354357 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8njr\" (UniqueName: \"kubernetes.io/projected/8b35f241-312f-4245-94f1-baf253d603bf-kube-api-access-b8njr\") pod \"8b35f241-312f-4245-94f1-baf253d603bf\" (UID: \"8b35f241-312f-4245-94f1-baf253d603bf\") " Dec 27 08:52:43 crc kubenswrapper[4934]: I1227 08:52:43.354803 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b35f241-312f-4245-94f1-baf253d603bf-utilities\") pod \"8b35f241-312f-4245-94f1-baf253d603bf\" (UID: \"8b35f241-312f-4245-94f1-baf253d603bf\") " Dec 27 08:52:43 crc kubenswrapper[4934]: I1227 08:52:43.354874 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b35f241-312f-4245-94f1-baf253d603bf-catalog-content\") pod \"8b35f241-312f-4245-94f1-baf253d603bf\" (UID: \"8b35f241-312f-4245-94f1-baf253d603bf\") " Dec 27 08:52:43 crc kubenswrapper[4934]: I1227 08:52:43.355542 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b35f241-312f-4245-94f1-baf253d603bf-utilities" (OuterVolumeSpecName: "utilities") pod "8b35f241-312f-4245-94f1-baf253d603bf" (UID: "8b35f241-312f-4245-94f1-baf253d603bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:52:43 crc kubenswrapper[4934]: I1227 08:52:43.360804 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b35f241-312f-4245-94f1-baf253d603bf-kube-api-access-b8njr" (OuterVolumeSpecName: "kube-api-access-b8njr") pod "8b35f241-312f-4245-94f1-baf253d603bf" (UID: "8b35f241-312f-4245-94f1-baf253d603bf"). InnerVolumeSpecName "kube-api-access-b8njr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:52:43 crc kubenswrapper[4934]: I1227 08:52:43.377015 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b35f241-312f-4245-94f1-baf253d603bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8b35f241-312f-4245-94f1-baf253d603bf" (UID: "8b35f241-312f-4245-94f1-baf253d603bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:52:43 crc kubenswrapper[4934]: I1227 08:52:43.458517 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8njr\" (UniqueName: \"kubernetes.io/projected/8b35f241-312f-4245-94f1-baf253d603bf-kube-api-access-b8njr\") on node \"crc\" DevicePath \"\"" Dec 27 08:52:43 crc kubenswrapper[4934]: I1227 08:52:43.458562 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b35f241-312f-4245-94f1-baf253d603bf-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 08:52:43 crc kubenswrapper[4934]: I1227 08:52:43.458577 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b35f241-312f-4245-94f1-baf253d603bf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 08:52:44 crc kubenswrapper[4934]: I1227 08:52:44.139028 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5m94h" Dec 27 08:52:44 crc kubenswrapper[4934]: I1227 08:52:44.165453 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5m94h"] Dec 27 08:52:44 crc kubenswrapper[4934]: I1227 08:52:44.175994 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5m94h"] Dec 27 08:52:45 crc kubenswrapper[4934]: I1227 08:52:45.482665 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b35f241-312f-4245-94f1-baf253d603bf" path="/var/lib/kubelet/pods/8b35f241-312f-4245-94f1-baf253d603bf/volumes" Dec 27 08:53:45 crc kubenswrapper[4934]: I1227 08:53:45.330485 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:53:45 crc kubenswrapper[4934]: I1227 08:53:45.331109 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:54:00 crc kubenswrapper[4934]: E1227 08:54:00.013328 4934 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.13:51578->38.102.83.13:43999: write tcp 38.102.83.13:51578->38.102.83.13:43999: write: broken pipe Dec 27 08:54:14 crc kubenswrapper[4934]: E1227 08:54:14.507274 4934 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.13:47274->38.102.83.13:43999: write tcp 38.102.83.13:47274->38.102.83.13:43999: write: broken pipe Dec 27 08:54:15 crc kubenswrapper[4934]: I1227 08:54:15.329755 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:54:15 crc kubenswrapper[4934]: I1227 08:54:15.330282 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:54:45 crc kubenswrapper[4934]: I1227 08:54:45.330834 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 08:54:45 crc kubenswrapper[4934]: I1227 08:54:45.331512 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 08:54:45 crc kubenswrapper[4934]: I1227 08:54:45.331566 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 08:54:45 crc kubenswrapper[4934]: I1227 08:54:45.332612 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 08:54:45 crc kubenswrapper[4934]: I1227 08:54:45.332772 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" gracePeriod=600 Dec 27 08:54:45 crc kubenswrapper[4934]: E1227 08:54:45.473767 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:54:45 crc kubenswrapper[4934]: I1227 08:54:45.596357 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84"} Dec 27 08:54:45 crc kubenswrapper[4934]: I1227 08:54:45.596315 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" exitCode=0 Dec 27 08:54:45 crc kubenswrapper[4934]: I1227 08:54:45.596638 4934 scope.go:117] "RemoveContainer" containerID="c468952548ae16cc22785c126991e9b5b77b35573edaa125b08c1c9d98f746ac" Dec 27 08:54:45 crc kubenswrapper[4934]: I1227 08:54:45.597522 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:54:45 crc kubenswrapper[4934]: E1227 08:54:45.597955 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:54:59 crc kubenswrapper[4934]: I1227 08:54:59.478130 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:54:59 crc kubenswrapper[4934]: E1227 08:54:59.479419 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:55:10 crc kubenswrapper[4934]: I1227 08:55:10.468402 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:55:10 crc kubenswrapper[4934]: E1227 08:55:10.470403 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:55:22 crc kubenswrapper[4934]: I1227 08:55:22.467444 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:55:22 crc kubenswrapper[4934]: E1227 08:55:22.468278 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:55:36 crc kubenswrapper[4934]: I1227 08:55:36.469131 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:55:36 crc kubenswrapper[4934]: E1227 08:55:36.470731 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:55:48 crc kubenswrapper[4934]: I1227 08:55:48.468318 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:55:48 crc kubenswrapper[4934]: E1227 08:55:48.469814 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:56:00 crc kubenswrapper[4934]: I1227 08:56:00.467575 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:56:00 crc kubenswrapper[4934]: E1227 08:56:00.468705 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:56:14 crc kubenswrapper[4934]: I1227 08:56:14.469027 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:56:14 crc kubenswrapper[4934]: E1227 08:56:14.470028 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:56:27 crc kubenswrapper[4934]: I1227 08:56:27.469461 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:56:27 crc kubenswrapper[4934]: E1227 08:56:27.470882 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:56:38 crc kubenswrapper[4934]: I1227 08:56:38.467885 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:56:38 crc kubenswrapper[4934]: E1227 08:56:38.468859 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:56:50 crc kubenswrapper[4934]: I1227 08:56:50.468720 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:56:50 crc kubenswrapper[4934]: E1227 08:56:50.469419 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:57:04 crc kubenswrapper[4934]: I1227 08:57:04.468506 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:57:04 crc kubenswrapper[4934]: E1227 08:57:04.469454 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:57:16 crc kubenswrapper[4934]: I1227 08:57:16.468733 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:57:16 crc kubenswrapper[4934]: E1227 08:57:16.469927 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:57:17 crc kubenswrapper[4934]: I1227 08:57:17.824721 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ls495"] Dec 27 08:57:17 crc kubenswrapper[4934]: E1227 08:57:17.825722 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b35f241-312f-4245-94f1-baf253d603bf" containerName="extract-utilities" Dec 27 08:57:17 crc kubenswrapper[4934]: I1227 08:57:17.825740 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b35f241-312f-4245-94f1-baf253d603bf" containerName="extract-utilities" Dec 27 08:57:17 crc kubenswrapper[4934]: E1227 08:57:17.825766 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b35f241-312f-4245-94f1-baf253d603bf" containerName="extract-content" Dec 27 08:57:17 crc kubenswrapper[4934]: I1227 08:57:17.825774 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b35f241-312f-4245-94f1-baf253d603bf" containerName="extract-content" Dec 27 08:57:17 crc kubenswrapper[4934]: E1227 08:57:17.825792 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b35f241-312f-4245-94f1-baf253d603bf" containerName="registry-server" Dec 27 08:57:17 crc kubenswrapper[4934]: I1227 08:57:17.825800 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b35f241-312f-4245-94f1-baf253d603bf" containerName="registry-server" Dec 27 08:57:17 crc kubenswrapper[4934]: I1227 08:57:17.826155 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b35f241-312f-4245-94f1-baf253d603bf" containerName="registry-server" Dec 27 08:57:17 crc kubenswrapper[4934]: I1227 08:57:17.828275 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ls495" Dec 27 08:57:17 crc kubenswrapper[4934]: I1227 08:57:17.858369 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ls495"] Dec 27 08:57:17 crc kubenswrapper[4934]: I1227 08:57:17.876594 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r46n5\" (UniqueName: \"kubernetes.io/projected/53a314da-1824-4849-b7e4-29b7a8b61a7f-kube-api-access-r46n5\") pod \"community-operators-ls495\" (UID: \"53a314da-1824-4849-b7e4-29b7a8b61a7f\") " pod="openshift-marketplace/community-operators-ls495" Dec 27 08:57:17 crc kubenswrapper[4934]: I1227 08:57:17.876782 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53a314da-1824-4849-b7e4-29b7a8b61a7f-utilities\") pod \"community-operators-ls495\" (UID: \"53a314da-1824-4849-b7e4-29b7a8b61a7f\") " pod="openshift-marketplace/community-operators-ls495" Dec 27 08:57:17 crc kubenswrapper[4934]: I1227 08:57:17.877796 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53a314da-1824-4849-b7e4-29b7a8b61a7f-catalog-content\") pod \"community-operators-ls495\" (UID: \"53a314da-1824-4849-b7e4-29b7a8b61a7f\") " pod="openshift-marketplace/community-operators-ls495" Dec 27 08:57:17 crc kubenswrapper[4934]: I1227 08:57:17.979617 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r46n5\" (UniqueName: \"kubernetes.io/projected/53a314da-1824-4849-b7e4-29b7a8b61a7f-kube-api-access-r46n5\") pod \"community-operators-ls495\" (UID: \"53a314da-1824-4849-b7e4-29b7a8b61a7f\") " pod="openshift-marketplace/community-operators-ls495" Dec 27 08:57:17 crc kubenswrapper[4934]: I1227 08:57:17.979709 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53a314da-1824-4849-b7e4-29b7a8b61a7f-utilities\") pod \"community-operators-ls495\" (UID: \"53a314da-1824-4849-b7e4-29b7a8b61a7f\") " pod="openshift-marketplace/community-operators-ls495" Dec 27 08:57:17 crc kubenswrapper[4934]: I1227 08:57:17.979797 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53a314da-1824-4849-b7e4-29b7a8b61a7f-catalog-content\") pod \"community-operators-ls495\" (UID: \"53a314da-1824-4849-b7e4-29b7a8b61a7f\") " pod="openshift-marketplace/community-operators-ls495" Dec 27 08:57:17 crc kubenswrapper[4934]: I1227 08:57:17.980278 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53a314da-1824-4849-b7e4-29b7a8b61a7f-catalog-content\") pod \"community-operators-ls495\" (UID: \"53a314da-1824-4849-b7e4-29b7a8b61a7f\") " pod="openshift-marketplace/community-operators-ls495" Dec 27 08:57:17 crc kubenswrapper[4934]: I1227 08:57:17.980729 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53a314da-1824-4849-b7e4-29b7a8b61a7f-utilities\") pod \"community-operators-ls495\" (UID: \"53a314da-1824-4849-b7e4-29b7a8b61a7f\") " pod="openshift-marketplace/community-operators-ls495" Dec 27 08:57:18 crc kubenswrapper[4934]: I1227 08:57:18.001705 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r46n5\" (UniqueName: \"kubernetes.io/projected/53a314da-1824-4849-b7e4-29b7a8b61a7f-kube-api-access-r46n5\") pod \"community-operators-ls495\" (UID: \"53a314da-1824-4849-b7e4-29b7a8b61a7f\") " pod="openshift-marketplace/community-operators-ls495" Dec 27 08:57:18 crc kubenswrapper[4934]: I1227 08:57:18.164882 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ls495" Dec 27 08:57:18 crc kubenswrapper[4934]: I1227 08:57:18.718986 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ls495"] Dec 27 08:57:18 crc kubenswrapper[4934]: W1227 08:57:18.726462 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53a314da_1824_4849_b7e4_29b7a8b61a7f.slice/crio-540880a626766d7e21487fad894939e5c41f34a3ed83d9b4584735f557fd4c87 WatchSource:0}: Error finding container 540880a626766d7e21487fad894939e5c41f34a3ed83d9b4584735f557fd4c87: Status 404 returned error can't find the container with id 540880a626766d7e21487fad894939e5c41f34a3ed83d9b4584735f557fd4c87 Dec 27 08:57:19 crc kubenswrapper[4934]: I1227 08:57:19.571821 4934 generic.go:334] "Generic (PLEG): container finished" podID="53a314da-1824-4849-b7e4-29b7a8b61a7f" containerID="2b7ddcb093719b076fa04f41f0c46d335dd74bdb9dabc5b66880d8122845b737" exitCode=0 Dec 27 08:57:19 crc kubenswrapper[4934]: I1227 08:57:19.571869 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ls495" event={"ID":"53a314da-1824-4849-b7e4-29b7a8b61a7f","Type":"ContainerDied","Data":"2b7ddcb093719b076fa04f41f0c46d335dd74bdb9dabc5b66880d8122845b737"} Dec 27 08:57:19 crc kubenswrapper[4934]: I1227 08:57:19.572338 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ls495" event={"ID":"53a314da-1824-4849-b7e4-29b7a8b61a7f","Type":"ContainerStarted","Data":"540880a626766d7e21487fad894939e5c41f34a3ed83d9b4584735f557fd4c87"} Dec 27 08:57:19 crc kubenswrapper[4934]: I1227 08:57:19.575267 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 27 08:57:21 crc kubenswrapper[4934]: I1227 08:57:21.599467 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ls495" event={"ID":"53a314da-1824-4849-b7e4-29b7a8b61a7f","Type":"ContainerStarted","Data":"f0b25a7ca3c3b3d96024999656655fad5c78902de6acb5a2f6cb486e9c523576"} Dec 27 08:57:22 crc kubenswrapper[4934]: I1227 08:57:22.615787 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ls495" event={"ID":"53a314da-1824-4849-b7e4-29b7a8b61a7f","Type":"ContainerDied","Data":"f0b25a7ca3c3b3d96024999656655fad5c78902de6acb5a2f6cb486e9c523576"} Dec 27 08:57:22 crc kubenswrapper[4934]: I1227 08:57:22.615666 4934 generic.go:334] "Generic (PLEG): container finished" podID="53a314da-1824-4849-b7e4-29b7a8b61a7f" containerID="f0b25a7ca3c3b3d96024999656655fad5c78902de6acb5a2f6cb486e9c523576" exitCode=0 Dec 27 08:57:23 crc kubenswrapper[4934]: I1227 08:57:23.628737 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ls495" event={"ID":"53a314da-1824-4849-b7e4-29b7a8b61a7f","Type":"ContainerStarted","Data":"86123ad9bf0c4e373bb04f8b619ec2f6266fde9e20c09406b0f3b0b9d6134502"} Dec 27 08:57:23 crc kubenswrapper[4934]: I1227 08:57:23.656059 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ls495" podStartSLOduration=3.007406911 podStartE2EDuration="6.656034627s" podCreationTimestamp="2025-12-27 08:57:17 +0000 UTC" firstStartedPulling="2025-12-27 08:57:19.574959191 +0000 UTC m=+4500.395399795" lastFinishedPulling="2025-12-27 08:57:23.223586917 +0000 UTC m=+4504.044027511" observedRunningTime="2025-12-27 08:57:23.647353692 +0000 UTC m=+4504.467794286" watchObservedRunningTime="2025-12-27 08:57:23.656034627 +0000 UTC m=+4504.476475221" Dec 27 08:57:28 crc kubenswrapper[4934]: I1227 08:57:28.165985 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ls495" Dec 27 08:57:28 crc kubenswrapper[4934]: I1227 08:57:28.166656 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ls495" Dec 27 08:57:28 crc kubenswrapper[4934]: I1227 08:57:28.240245 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ls495" Dec 27 08:57:28 crc kubenswrapper[4934]: I1227 08:57:28.738697 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ls495" Dec 27 08:57:28 crc kubenswrapper[4934]: I1227 08:57:28.810416 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ls495"] Dec 27 08:57:29 crc kubenswrapper[4934]: I1227 08:57:29.479651 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:57:29 crc kubenswrapper[4934]: E1227 08:57:29.479969 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:57:30 crc kubenswrapper[4934]: I1227 08:57:30.717365 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ls495" podUID="53a314da-1824-4849-b7e4-29b7a8b61a7f" containerName="registry-server" containerID="cri-o://86123ad9bf0c4e373bb04f8b619ec2f6266fde9e20c09406b0f3b0b9d6134502" gracePeriod=2 Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.281819 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ls495" Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.414780 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r46n5\" (UniqueName: \"kubernetes.io/projected/53a314da-1824-4849-b7e4-29b7a8b61a7f-kube-api-access-r46n5\") pod \"53a314da-1824-4849-b7e4-29b7a8b61a7f\" (UID: \"53a314da-1824-4849-b7e4-29b7a8b61a7f\") " Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.414941 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53a314da-1824-4849-b7e4-29b7a8b61a7f-utilities\") pod \"53a314da-1824-4849-b7e4-29b7a8b61a7f\" (UID: \"53a314da-1824-4849-b7e4-29b7a8b61a7f\") " Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.415110 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53a314da-1824-4849-b7e4-29b7a8b61a7f-catalog-content\") pod \"53a314da-1824-4849-b7e4-29b7a8b61a7f\" (UID: \"53a314da-1824-4849-b7e4-29b7a8b61a7f\") " Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.415691 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53a314da-1824-4849-b7e4-29b7a8b61a7f-utilities" (OuterVolumeSpecName: "utilities") pod "53a314da-1824-4849-b7e4-29b7a8b61a7f" (UID: "53a314da-1824-4849-b7e4-29b7a8b61a7f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.416334 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53a314da-1824-4849-b7e4-29b7a8b61a7f-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.423190 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53a314da-1824-4849-b7e4-29b7a8b61a7f-kube-api-access-r46n5" (OuterVolumeSpecName: "kube-api-access-r46n5") pod "53a314da-1824-4849-b7e4-29b7a8b61a7f" (UID: "53a314da-1824-4849-b7e4-29b7a8b61a7f"). InnerVolumeSpecName "kube-api-access-r46n5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.480480 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53a314da-1824-4849-b7e4-29b7a8b61a7f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "53a314da-1824-4849-b7e4-29b7a8b61a7f" (UID: "53a314da-1824-4849-b7e4-29b7a8b61a7f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.519260 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r46n5\" (UniqueName: \"kubernetes.io/projected/53a314da-1824-4849-b7e4-29b7a8b61a7f-kube-api-access-r46n5\") on node \"crc\" DevicePath \"\"" Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.519331 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53a314da-1824-4849-b7e4-29b7a8b61a7f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.738983 4934 generic.go:334] "Generic (PLEG): container finished" podID="53a314da-1824-4849-b7e4-29b7a8b61a7f" containerID="86123ad9bf0c4e373bb04f8b619ec2f6266fde9e20c09406b0f3b0b9d6134502" exitCode=0 Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.739031 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ls495" Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.739051 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ls495" event={"ID":"53a314da-1824-4849-b7e4-29b7a8b61a7f","Type":"ContainerDied","Data":"86123ad9bf0c4e373bb04f8b619ec2f6266fde9e20c09406b0f3b0b9d6134502"} Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.740363 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ls495" event={"ID":"53a314da-1824-4849-b7e4-29b7a8b61a7f","Type":"ContainerDied","Data":"540880a626766d7e21487fad894939e5c41f34a3ed83d9b4584735f557fd4c87"} Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.740408 4934 scope.go:117] "RemoveContainer" containerID="86123ad9bf0c4e373bb04f8b619ec2f6266fde9e20c09406b0f3b0b9d6134502" Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.772609 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ls495"] Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.783609 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ls495"] Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.810443 4934 scope.go:117] "RemoveContainer" containerID="f0b25a7ca3c3b3d96024999656655fad5c78902de6acb5a2f6cb486e9c523576" Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.845520 4934 scope.go:117] "RemoveContainer" containerID="2b7ddcb093719b076fa04f41f0c46d335dd74bdb9dabc5b66880d8122845b737" Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.918804 4934 scope.go:117] "RemoveContainer" containerID="86123ad9bf0c4e373bb04f8b619ec2f6266fde9e20c09406b0f3b0b9d6134502" Dec 27 08:57:31 crc kubenswrapper[4934]: E1227 08:57:31.919249 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86123ad9bf0c4e373bb04f8b619ec2f6266fde9e20c09406b0f3b0b9d6134502\": container with ID starting with 86123ad9bf0c4e373bb04f8b619ec2f6266fde9e20c09406b0f3b0b9d6134502 not found: ID does not exist" containerID="86123ad9bf0c4e373bb04f8b619ec2f6266fde9e20c09406b0f3b0b9d6134502" Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.919283 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86123ad9bf0c4e373bb04f8b619ec2f6266fde9e20c09406b0f3b0b9d6134502"} err="failed to get container status \"86123ad9bf0c4e373bb04f8b619ec2f6266fde9e20c09406b0f3b0b9d6134502\": rpc error: code = NotFound desc = could not find container \"86123ad9bf0c4e373bb04f8b619ec2f6266fde9e20c09406b0f3b0b9d6134502\": container with ID starting with 86123ad9bf0c4e373bb04f8b619ec2f6266fde9e20c09406b0f3b0b9d6134502 not found: ID does not exist" Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.919310 4934 scope.go:117] "RemoveContainer" containerID="f0b25a7ca3c3b3d96024999656655fad5c78902de6acb5a2f6cb486e9c523576" Dec 27 08:57:31 crc kubenswrapper[4934]: E1227 08:57:31.919656 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0b25a7ca3c3b3d96024999656655fad5c78902de6acb5a2f6cb486e9c523576\": container with ID starting with f0b25a7ca3c3b3d96024999656655fad5c78902de6acb5a2f6cb486e9c523576 not found: ID does not exist" containerID="f0b25a7ca3c3b3d96024999656655fad5c78902de6acb5a2f6cb486e9c523576" Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.919713 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0b25a7ca3c3b3d96024999656655fad5c78902de6acb5a2f6cb486e9c523576"} err="failed to get container status \"f0b25a7ca3c3b3d96024999656655fad5c78902de6acb5a2f6cb486e9c523576\": rpc error: code = NotFound desc = could not find container \"f0b25a7ca3c3b3d96024999656655fad5c78902de6acb5a2f6cb486e9c523576\": container with ID starting with f0b25a7ca3c3b3d96024999656655fad5c78902de6acb5a2f6cb486e9c523576 not found: ID does not exist" Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.919746 4934 scope.go:117] "RemoveContainer" containerID="2b7ddcb093719b076fa04f41f0c46d335dd74bdb9dabc5b66880d8122845b737" Dec 27 08:57:31 crc kubenswrapper[4934]: E1227 08:57:31.920101 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b7ddcb093719b076fa04f41f0c46d335dd74bdb9dabc5b66880d8122845b737\": container with ID starting with 2b7ddcb093719b076fa04f41f0c46d335dd74bdb9dabc5b66880d8122845b737 not found: ID does not exist" containerID="2b7ddcb093719b076fa04f41f0c46d335dd74bdb9dabc5b66880d8122845b737" Dec 27 08:57:31 crc kubenswrapper[4934]: I1227 08:57:31.920131 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b7ddcb093719b076fa04f41f0c46d335dd74bdb9dabc5b66880d8122845b737"} err="failed to get container status \"2b7ddcb093719b076fa04f41f0c46d335dd74bdb9dabc5b66880d8122845b737\": rpc error: code = NotFound desc = could not find container \"2b7ddcb093719b076fa04f41f0c46d335dd74bdb9dabc5b66880d8122845b737\": container with ID starting with 2b7ddcb093719b076fa04f41f0c46d335dd74bdb9dabc5b66880d8122845b737 not found: ID does not exist" Dec 27 08:57:33 crc kubenswrapper[4934]: I1227 08:57:33.489120 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53a314da-1824-4849-b7e4-29b7a8b61a7f" path="/var/lib/kubelet/pods/53a314da-1824-4849-b7e4-29b7a8b61a7f/volumes" Dec 27 08:57:39 crc kubenswrapper[4934]: I1227 08:57:39.952617 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-4jwwj" podUID="ea01c7df-c5aa-4c86-800b-b6aebafeb8fd" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 08:57:39 crc kubenswrapper[4934]: I1227 08:57:39.953826 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-4jwwj" podUID="ea01c7df-c5aa-4c86-800b-b6aebafeb8fd" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 08:57:44 crc kubenswrapper[4934]: I1227 08:57:44.467885 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:57:44 crc kubenswrapper[4934]: E1227 08:57:44.468813 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:57:58 crc kubenswrapper[4934]: I1227 08:57:58.468563 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:57:58 crc kubenswrapper[4934]: E1227 08:57:58.469467 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:58:09 crc kubenswrapper[4934]: I1227 08:58:09.476556 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:58:09 crc kubenswrapper[4934]: E1227 08:58:09.478617 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:58:20 crc kubenswrapper[4934]: I1227 08:58:20.468096 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:58:20 crc kubenswrapper[4934]: E1227 08:58:20.468948 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:58:35 crc kubenswrapper[4934]: I1227 08:58:35.468275 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:58:35 crc kubenswrapper[4934]: E1227 08:58:35.469413 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:58:48 crc kubenswrapper[4934]: I1227 08:58:48.468132 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:58:48 crc kubenswrapper[4934]: E1227 08:58:48.468887 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:59:03 crc kubenswrapper[4934]: I1227 08:59:03.467478 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:59:03 crc kubenswrapper[4934]: E1227 08:59:03.468512 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:59:17 crc kubenswrapper[4934]: I1227 08:59:17.468431 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:59:17 crc kubenswrapper[4934]: E1227 08:59:17.469824 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:59:29 crc kubenswrapper[4934]: I1227 08:59:29.773679 4934 scope.go:117] "RemoveContainer" containerID="10bd634712e3b7a6815ad769496b68ee138254d5d4f936f4b5f44669630a4a48" Dec 27 08:59:29 crc kubenswrapper[4934]: I1227 08:59:29.830258 4934 scope.go:117] "RemoveContainer" containerID="ba8c3f5e5dec3fb1e4b72defdf9f150348c79064c27a61eb5ff8514aea6b71d0" Dec 27 08:59:29 crc kubenswrapper[4934]: I1227 08:59:29.898049 4934 scope.go:117] "RemoveContainer" containerID="96e05ae1a9be16e7dd9fae3b841864e14dfeb187c0653e01490f67e3b32e9e2d" Dec 27 08:59:32 crc kubenswrapper[4934]: I1227 08:59:32.468600 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:59:32 crc kubenswrapper[4934]: E1227 08:59:32.470163 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 08:59:46 crc kubenswrapper[4934]: I1227 08:59:46.467731 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 08:59:47 crc kubenswrapper[4934]: I1227 08:59:47.622865 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"8224ba78a25faf72ee4cb82888cba291e56d3f159515c14eec31cad4a6c0f282"} Dec 27 09:00:00 crc kubenswrapper[4934]: I1227 09:00:00.183069 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447100-tthxq"] Dec 27 09:00:00 crc kubenswrapper[4934]: E1227 09:00:00.184475 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53a314da-1824-4849-b7e4-29b7a8b61a7f" containerName="extract-utilities" Dec 27 09:00:00 crc kubenswrapper[4934]: I1227 09:00:00.184500 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="53a314da-1824-4849-b7e4-29b7a8b61a7f" containerName="extract-utilities" Dec 27 09:00:00 crc kubenswrapper[4934]: E1227 09:00:00.184574 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53a314da-1824-4849-b7e4-29b7a8b61a7f" containerName="extract-content" Dec 27 09:00:00 crc kubenswrapper[4934]: I1227 09:00:00.184586 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="53a314da-1824-4849-b7e4-29b7a8b61a7f" containerName="extract-content" Dec 27 09:00:00 crc kubenswrapper[4934]: E1227 09:00:00.184613 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53a314da-1824-4849-b7e4-29b7a8b61a7f" containerName="registry-server" Dec 27 09:00:00 crc kubenswrapper[4934]: I1227 09:00:00.184622 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="53a314da-1824-4849-b7e4-29b7a8b61a7f" containerName="registry-server" Dec 27 09:00:00 crc kubenswrapper[4934]: I1227 09:00:00.185030 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="53a314da-1824-4849-b7e4-29b7a8b61a7f" containerName="registry-server" Dec 27 09:00:00 crc kubenswrapper[4934]: I1227 09:00:00.186511 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447100-tthxq" Dec 27 09:00:00 crc kubenswrapper[4934]: I1227 09:00:00.189828 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 27 09:00:00 crc kubenswrapper[4934]: I1227 09:00:00.196250 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 27 09:00:00 crc kubenswrapper[4934]: I1227 09:00:00.198161 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447100-tthxq"] Dec 27 09:00:00 crc kubenswrapper[4934]: I1227 09:00:00.274452 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6qmt\" (UniqueName: \"kubernetes.io/projected/b5e678b1-0f0b-4319-a146-48b557e9c107-kube-api-access-j6qmt\") pod \"collect-profiles-29447100-tthxq\" (UID: \"b5e678b1-0f0b-4319-a146-48b557e9c107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447100-tthxq" Dec 27 09:00:00 crc kubenswrapper[4934]: I1227 09:00:00.274602 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b5e678b1-0f0b-4319-a146-48b557e9c107-secret-volume\") pod \"collect-profiles-29447100-tthxq\" (UID: \"b5e678b1-0f0b-4319-a146-48b557e9c107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447100-tthxq" Dec 27 09:00:00 crc kubenswrapper[4934]: I1227 09:00:00.274651 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b5e678b1-0f0b-4319-a146-48b557e9c107-config-volume\") pod \"collect-profiles-29447100-tthxq\" (UID: \"b5e678b1-0f0b-4319-a146-48b557e9c107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447100-tthxq" Dec 27 09:00:00 crc kubenswrapper[4934]: I1227 09:00:00.377039 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b5e678b1-0f0b-4319-a146-48b557e9c107-secret-volume\") pod \"collect-profiles-29447100-tthxq\" (UID: \"b5e678b1-0f0b-4319-a146-48b557e9c107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447100-tthxq" Dec 27 09:00:00 crc kubenswrapper[4934]: I1227 09:00:00.377167 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b5e678b1-0f0b-4319-a146-48b557e9c107-config-volume\") pod \"collect-profiles-29447100-tthxq\" (UID: \"b5e678b1-0f0b-4319-a146-48b557e9c107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447100-tthxq" Dec 27 09:00:00 crc kubenswrapper[4934]: I1227 09:00:00.377341 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6qmt\" (UniqueName: \"kubernetes.io/projected/b5e678b1-0f0b-4319-a146-48b557e9c107-kube-api-access-j6qmt\") pod \"collect-profiles-29447100-tthxq\" (UID: \"b5e678b1-0f0b-4319-a146-48b557e9c107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447100-tthxq" Dec 27 09:00:00 crc kubenswrapper[4934]: I1227 09:00:00.377898 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b5e678b1-0f0b-4319-a146-48b557e9c107-config-volume\") pod \"collect-profiles-29447100-tthxq\" (UID: \"b5e678b1-0f0b-4319-a146-48b557e9c107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447100-tthxq" Dec 27 09:00:00 crc kubenswrapper[4934]: I1227 09:00:00.384577 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b5e678b1-0f0b-4319-a146-48b557e9c107-secret-volume\") pod \"collect-profiles-29447100-tthxq\" (UID: \"b5e678b1-0f0b-4319-a146-48b557e9c107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447100-tthxq" Dec 27 09:00:00 crc kubenswrapper[4934]: I1227 09:00:00.395264 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6qmt\" (UniqueName: \"kubernetes.io/projected/b5e678b1-0f0b-4319-a146-48b557e9c107-kube-api-access-j6qmt\") pod \"collect-profiles-29447100-tthxq\" (UID: \"b5e678b1-0f0b-4319-a146-48b557e9c107\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447100-tthxq" Dec 27 09:00:00 crc kubenswrapper[4934]: I1227 09:00:00.513515 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447100-tthxq" Dec 27 09:00:01 crc kubenswrapper[4934]: I1227 09:00:01.007772 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447100-tthxq"] Dec 27 09:00:01 crc kubenswrapper[4934]: W1227 09:00:01.012139 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5e678b1_0f0b_4319_a146_48b557e9c107.slice/crio-e1f793edd2300038b6cb75c8cd5043a00cd54ab3b44343ed709e736936b1b0ea WatchSource:0}: Error finding container e1f793edd2300038b6cb75c8cd5043a00cd54ab3b44343ed709e736936b1b0ea: Status 404 returned error can't find the container with id e1f793edd2300038b6cb75c8cd5043a00cd54ab3b44343ed709e736936b1b0ea Dec 27 09:00:01 crc kubenswrapper[4934]: I1227 09:00:01.809228 4934 generic.go:334] "Generic (PLEG): container finished" podID="b5e678b1-0f0b-4319-a146-48b557e9c107" containerID="2c6f2f530caf0dc3a599e1c5e8761d0f6d16f2bca859fc456a39c547bfbde8aa" exitCode=0 Dec 27 09:00:01 crc kubenswrapper[4934]: I1227 09:00:01.809303 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447100-tthxq" event={"ID":"b5e678b1-0f0b-4319-a146-48b557e9c107","Type":"ContainerDied","Data":"2c6f2f530caf0dc3a599e1c5e8761d0f6d16f2bca859fc456a39c547bfbde8aa"} Dec 27 09:00:01 crc kubenswrapper[4934]: I1227 09:00:01.809623 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447100-tthxq" event={"ID":"b5e678b1-0f0b-4319-a146-48b557e9c107","Type":"ContainerStarted","Data":"e1f793edd2300038b6cb75c8cd5043a00cd54ab3b44343ed709e736936b1b0ea"} Dec 27 09:00:03 crc kubenswrapper[4934]: I1227 09:00:03.231831 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447100-tthxq" Dec 27 09:00:03 crc kubenswrapper[4934]: I1227 09:00:03.381247 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b5e678b1-0f0b-4319-a146-48b557e9c107-config-volume\") pod \"b5e678b1-0f0b-4319-a146-48b557e9c107\" (UID: \"b5e678b1-0f0b-4319-a146-48b557e9c107\") " Dec 27 09:00:03 crc kubenswrapper[4934]: I1227 09:00:03.381390 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6qmt\" (UniqueName: \"kubernetes.io/projected/b5e678b1-0f0b-4319-a146-48b557e9c107-kube-api-access-j6qmt\") pod \"b5e678b1-0f0b-4319-a146-48b557e9c107\" (UID: \"b5e678b1-0f0b-4319-a146-48b557e9c107\") " Dec 27 09:00:03 crc kubenswrapper[4934]: I1227 09:00:03.381427 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b5e678b1-0f0b-4319-a146-48b557e9c107-secret-volume\") pod \"b5e678b1-0f0b-4319-a146-48b557e9c107\" (UID: \"b5e678b1-0f0b-4319-a146-48b557e9c107\") " Dec 27 09:00:03 crc kubenswrapper[4934]: I1227 09:00:03.382535 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5e678b1-0f0b-4319-a146-48b557e9c107-config-volume" (OuterVolumeSpecName: "config-volume") pod "b5e678b1-0f0b-4319-a146-48b557e9c107" (UID: "b5e678b1-0f0b-4319-a146-48b557e9c107"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 09:00:03 crc kubenswrapper[4934]: I1227 09:00:03.387949 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5e678b1-0f0b-4319-a146-48b557e9c107-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b5e678b1-0f0b-4319-a146-48b557e9c107" (UID: "b5e678b1-0f0b-4319-a146-48b557e9c107"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 09:00:03 crc kubenswrapper[4934]: I1227 09:00:03.405446 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5e678b1-0f0b-4319-a146-48b557e9c107-kube-api-access-j6qmt" (OuterVolumeSpecName: "kube-api-access-j6qmt") pod "b5e678b1-0f0b-4319-a146-48b557e9c107" (UID: "b5e678b1-0f0b-4319-a146-48b557e9c107"). InnerVolumeSpecName "kube-api-access-j6qmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:00:03 crc kubenswrapper[4934]: I1227 09:00:03.485065 4934 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b5e678b1-0f0b-4319-a146-48b557e9c107-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 27 09:00:03 crc kubenswrapper[4934]: I1227 09:00:03.485110 4934 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b5e678b1-0f0b-4319-a146-48b557e9c107-config-volume\") on node \"crc\" DevicePath \"\"" Dec 27 09:00:03 crc kubenswrapper[4934]: I1227 09:00:03.485123 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6qmt\" (UniqueName: \"kubernetes.io/projected/b5e678b1-0f0b-4319-a146-48b557e9c107-kube-api-access-j6qmt\") on node \"crc\" DevicePath \"\"" Dec 27 09:00:03 crc kubenswrapper[4934]: I1227 09:00:03.839794 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447100-tthxq" event={"ID":"b5e678b1-0f0b-4319-a146-48b557e9c107","Type":"ContainerDied","Data":"e1f793edd2300038b6cb75c8cd5043a00cd54ab3b44343ed709e736936b1b0ea"} Dec 27 09:00:03 crc kubenswrapper[4934]: I1227 09:00:03.840029 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1f793edd2300038b6cb75c8cd5043a00cd54ab3b44343ed709e736936b1b0ea" Dec 27 09:00:03 crc kubenswrapper[4934]: I1227 09:00:03.840098 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447100-tthxq" Dec 27 09:00:04 crc kubenswrapper[4934]: I1227 09:00:04.316614 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c"] Dec 27 09:00:04 crc kubenswrapper[4934]: I1227 09:00:04.329724 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447055-vg29c"] Dec 27 09:00:05 crc kubenswrapper[4934]: I1227 09:00:05.483735 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1457a36-6d7f-4770-bca1-d7bc5919bb4c" path="/var/lib/kubelet/pods/d1457a36-6d7f-4770-bca1-d7bc5919bb4c/volumes" Dec 27 09:00:29 crc kubenswrapper[4934]: I1227 09:00:29.982010 4934 scope.go:117] "RemoveContainer" containerID="d6f67af5a2fc359f79dffaecc06c0e92449e263cd574162ecfcec72d08c18736" Dec 27 09:00:31 crc kubenswrapper[4934]: E1227 09:00:31.376779 4934 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.13:54386->38.102.83.13:43999: read tcp 38.102.83.13:54386->38.102.83.13:43999: read: connection reset by peer Dec 27 09:00:33 crc kubenswrapper[4934]: I1227 09:00:33.048327 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nhkf7"] Dec 27 09:00:33 crc kubenswrapper[4934]: E1227 09:00:33.049124 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5e678b1-0f0b-4319-a146-48b557e9c107" containerName="collect-profiles" Dec 27 09:00:33 crc kubenswrapper[4934]: I1227 09:00:33.049138 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5e678b1-0f0b-4319-a146-48b557e9c107" containerName="collect-profiles" Dec 27 09:00:33 crc kubenswrapper[4934]: I1227 09:00:33.049388 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5e678b1-0f0b-4319-a146-48b557e9c107" containerName="collect-profiles" Dec 27 09:00:33 crc kubenswrapper[4934]: I1227 09:00:33.051286 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nhkf7" Dec 27 09:00:33 crc kubenswrapper[4934]: I1227 09:00:33.062475 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nhkf7"] Dec 27 09:00:33 crc kubenswrapper[4934]: I1227 09:00:33.150021 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4080093-7ee5-483e-9422-ec3837273b04-utilities\") pod \"certified-operators-nhkf7\" (UID: \"d4080093-7ee5-483e-9422-ec3837273b04\") " pod="openshift-marketplace/certified-operators-nhkf7" Dec 27 09:00:33 crc kubenswrapper[4934]: I1227 09:00:33.150575 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlwth\" (UniqueName: \"kubernetes.io/projected/d4080093-7ee5-483e-9422-ec3837273b04-kube-api-access-vlwth\") pod \"certified-operators-nhkf7\" (UID: \"d4080093-7ee5-483e-9422-ec3837273b04\") " pod="openshift-marketplace/certified-operators-nhkf7" Dec 27 09:00:33 crc kubenswrapper[4934]: I1227 09:00:33.150886 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4080093-7ee5-483e-9422-ec3837273b04-catalog-content\") pod \"certified-operators-nhkf7\" (UID: \"d4080093-7ee5-483e-9422-ec3837273b04\") " pod="openshift-marketplace/certified-operators-nhkf7" Dec 27 09:00:33 crc kubenswrapper[4934]: I1227 09:00:33.252808 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4080093-7ee5-483e-9422-ec3837273b04-utilities\") pod \"certified-operators-nhkf7\" (UID: \"d4080093-7ee5-483e-9422-ec3837273b04\") " pod="openshift-marketplace/certified-operators-nhkf7" Dec 27 09:00:33 crc kubenswrapper[4934]: I1227 09:00:33.252861 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlwth\" (UniqueName: \"kubernetes.io/projected/d4080093-7ee5-483e-9422-ec3837273b04-kube-api-access-vlwth\") pod \"certified-operators-nhkf7\" (UID: \"d4080093-7ee5-483e-9422-ec3837273b04\") " pod="openshift-marketplace/certified-operators-nhkf7" Dec 27 09:00:33 crc kubenswrapper[4934]: I1227 09:00:33.252972 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4080093-7ee5-483e-9422-ec3837273b04-catalog-content\") pod \"certified-operators-nhkf7\" (UID: \"d4080093-7ee5-483e-9422-ec3837273b04\") " pod="openshift-marketplace/certified-operators-nhkf7" Dec 27 09:00:33 crc kubenswrapper[4934]: I1227 09:00:33.253295 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4080093-7ee5-483e-9422-ec3837273b04-utilities\") pod \"certified-operators-nhkf7\" (UID: \"d4080093-7ee5-483e-9422-ec3837273b04\") " pod="openshift-marketplace/certified-operators-nhkf7" Dec 27 09:00:33 crc kubenswrapper[4934]: I1227 09:00:33.253366 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4080093-7ee5-483e-9422-ec3837273b04-catalog-content\") pod \"certified-operators-nhkf7\" (UID: \"d4080093-7ee5-483e-9422-ec3837273b04\") " pod="openshift-marketplace/certified-operators-nhkf7" Dec 27 09:00:33 crc kubenswrapper[4934]: I1227 09:00:33.739463 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlwth\" (UniqueName: \"kubernetes.io/projected/d4080093-7ee5-483e-9422-ec3837273b04-kube-api-access-vlwth\") pod \"certified-operators-nhkf7\" (UID: \"d4080093-7ee5-483e-9422-ec3837273b04\") " pod="openshift-marketplace/certified-operators-nhkf7" Dec 27 09:00:33 crc kubenswrapper[4934]: I1227 09:00:33.976452 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nhkf7" Dec 27 09:00:34 crc kubenswrapper[4934]: I1227 09:00:34.527576 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nhkf7"] Dec 27 09:00:35 crc kubenswrapper[4934]: I1227 09:00:35.211128 4934 generic.go:334] "Generic (PLEG): container finished" podID="d4080093-7ee5-483e-9422-ec3837273b04" containerID="67d26929fe9646ad72855b9fddb3d99b10353382c7b8a204f41ed68b92f2f8c7" exitCode=0 Dec 27 09:00:35 crc kubenswrapper[4934]: I1227 09:00:35.211265 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhkf7" event={"ID":"d4080093-7ee5-483e-9422-ec3837273b04","Type":"ContainerDied","Data":"67d26929fe9646ad72855b9fddb3d99b10353382c7b8a204f41ed68b92f2f8c7"} Dec 27 09:00:35 crc kubenswrapper[4934]: I1227 09:00:35.212076 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhkf7" event={"ID":"d4080093-7ee5-483e-9422-ec3837273b04","Type":"ContainerStarted","Data":"1644d1d17d939fa3b6e05a54de4b6297f9c0079cf0a12cb221413b3a333acd84"} Dec 27 09:00:37 crc kubenswrapper[4934]: I1227 09:00:37.238589 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhkf7" event={"ID":"d4080093-7ee5-483e-9422-ec3837273b04","Type":"ContainerStarted","Data":"ec933e62c05f27a1e3022793b1c2076f671ab36e8f9f294262afe4cc19454f34"} Dec 27 09:00:38 crc kubenswrapper[4934]: I1227 09:00:38.251394 4934 generic.go:334] "Generic (PLEG): container finished" podID="d4080093-7ee5-483e-9422-ec3837273b04" containerID="ec933e62c05f27a1e3022793b1c2076f671ab36e8f9f294262afe4cc19454f34" exitCode=0 Dec 27 09:00:38 crc kubenswrapper[4934]: I1227 09:00:38.251587 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhkf7" event={"ID":"d4080093-7ee5-483e-9422-ec3837273b04","Type":"ContainerDied","Data":"ec933e62c05f27a1e3022793b1c2076f671ab36e8f9f294262afe4cc19454f34"} Dec 27 09:00:39 crc kubenswrapper[4934]: I1227 09:00:39.267563 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhkf7" event={"ID":"d4080093-7ee5-483e-9422-ec3837273b04","Type":"ContainerStarted","Data":"720bbcf8a36d0e426e486808f4ff8caa0ef5a76a53ba59cd2e6faf96e9173f19"} Dec 27 09:00:39 crc kubenswrapper[4934]: I1227 09:00:39.293306 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nhkf7" podStartSLOduration=2.846463511 podStartE2EDuration="6.293281064s" podCreationTimestamp="2025-12-27 09:00:33 +0000 UTC" firstStartedPulling="2025-12-27 09:00:35.214458874 +0000 UTC m=+4696.034899468" lastFinishedPulling="2025-12-27 09:00:38.661276427 +0000 UTC m=+4699.481717021" observedRunningTime="2025-12-27 09:00:39.291198282 +0000 UTC m=+4700.111638946" watchObservedRunningTime="2025-12-27 09:00:39.293281064 +0000 UTC m=+4700.113721688" Dec 27 09:00:43 crc kubenswrapper[4934]: I1227 09:00:43.977513 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nhkf7" Dec 27 09:00:43 crc kubenswrapper[4934]: I1227 09:00:43.978272 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nhkf7" Dec 27 09:00:44 crc kubenswrapper[4934]: I1227 09:00:44.043305 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nhkf7" Dec 27 09:00:45 crc kubenswrapper[4934]: I1227 09:00:45.422403 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nhkf7" Dec 27 09:00:45 crc kubenswrapper[4934]: I1227 09:00:45.504075 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nhkf7"] Dec 27 09:00:46 crc kubenswrapper[4934]: I1227 09:00:46.346357 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nhkf7" podUID="d4080093-7ee5-483e-9422-ec3837273b04" containerName="registry-server" containerID="cri-o://720bbcf8a36d0e426e486808f4ff8caa0ef5a76a53ba59cd2e6faf96e9173f19" gracePeriod=2 Dec 27 09:00:47 crc kubenswrapper[4934]: I1227 09:00:47.357831 4934 generic.go:334] "Generic (PLEG): container finished" podID="d4080093-7ee5-483e-9422-ec3837273b04" containerID="720bbcf8a36d0e426e486808f4ff8caa0ef5a76a53ba59cd2e6faf96e9173f19" exitCode=0 Dec 27 09:00:47 crc kubenswrapper[4934]: I1227 09:00:47.357880 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhkf7" event={"ID":"d4080093-7ee5-483e-9422-ec3837273b04","Type":"ContainerDied","Data":"720bbcf8a36d0e426e486808f4ff8caa0ef5a76a53ba59cd2e6faf96e9173f19"} Dec 27 09:00:47 crc kubenswrapper[4934]: I1227 09:00:47.594961 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nhkf7" Dec 27 09:00:47 crc kubenswrapper[4934]: I1227 09:00:47.733690 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4080093-7ee5-483e-9422-ec3837273b04-catalog-content\") pod \"d4080093-7ee5-483e-9422-ec3837273b04\" (UID: \"d4080093-7ee5-483e-9422-ec3837273b04\") " Dec 27 09:00:47 crc kubenswrapper[4934]: I1227 09:00:47.733924 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4080093-7ee5-483e-9422-ec3837273b04-utilities\") pod \"d4080093-7ee5-483e-9422-ec3837273b04\" (UID: \"d4080093-7ee5-483e-9422-ec3837273b04\") " Dec 27 09:00:47 crc kubenswrapper[4934]: I1227 09:00:47.734036 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlwth\" (UniqueName: \"kubernetes.io/projected/d4080093-7ee5-483e-9422-ec3837273b04-kube-api-access-vlwth\") pod \"d4080093-7ee5-483e-9422-ec3837273b04\" (UID: \"d4080093-7ee5-483e-9422-ec3837273b04\") " Dec 27 09:00:47 crc kubenswrapper[4934]: I1227 09:00:47.734988 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4080093-7ee5-483e-9422-ec3837273b04-utilities" (OuterVolumeSpecName: "utilities") pod "d4080093-7ee5-483e-9422-ec3837273b04" (UID: "d4080093-7ee5-483e-9422-ec3837273b04"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:00:47 crc kubenswrapper[4934]: I1227 09:00:47.743147 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4080093-7ee5-483e-9422-ec3837273b04-kube-api-access-vlwth" (OuterVolumeSpecName: "kube-api-access-vlwth") pod "d4080093-7ee5-483e-9422-ec3837273b04" (UID: "d4080093-7ee5-483e-9422-ec3837273b04"). InnerVolumeSpecName "kube-api-access-vlwth". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:00:47 crc kubenswrapper[4934]: I1227 09:00:47.803504 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4080093-7ee5-483e-9422-ec3837273b04-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4080093-7ee5-483e-9422-ec3837273b04" (UID: "d4080093-7ee5-483e-9422-ec3837273b04"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:00:47 crc kubenswrapper[4934]: I1227 09:00:47.837417 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4080093-7ee5-483e-9422-ec3837273b04-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 09:00:47 crc kubenswrapper[4934]: I1227 09:00:47.837456 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlwth\" (UniqueName: \"kubernetes.io/projected/d4080093-7ee5-483e-9422-ec3837273b04-kube-api-access-vlwth\") on node \"crc\" DevicePath \"\"" Dec 27 09:00:47 crc kubenswrapper[4934]: I1227 09:00:47.837470 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4080093-7ee5-483e-9422-ec3837273b04-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 09:00:48 crc kubenswrapper[4934]: I1227 09:00:48.378204 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nhkf7" event={"ID":"d4080093-7ee5-483e-9422-ec3837273b04","Type":"ContainerDied","Data":"1644d1d17d939fa3b6e05a54de4b6297f9c0079cf0a12cb221413b3a333acd84"} Dec 27 09:00:48 crc kubenswrapper[4934]: I1227 09:00:48.378569 4934 scope.go:117] "RemoveContainer" containerID="720bbcf8a36d0e426e486808f4ff8caa0ef5a76a53ba59cd2e6faf96e9173f19" Dec 27 09:00:48 crc kubenswrapper[4934]: I1227 09:00:48.378313 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nhkf7" Dec 27 09:00:48 crc kubenswrapper[4934]: I1227 09:00:48.408856 4934 scope.go:117] "RemoveContainer" containerID="ec933e62c05f27a1e3022793b1c2076f671ab36e8f9f294262afe4cc19454f34" Dec 27 09:00:48 crc kubenswrapper[4934]: I1227 09:00:48.443358 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nhkf7"] Dec 27 09:00:48 crc kubenswrapper[4934]: I1227 09:00:48.459403 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nhkf7"] Dec 27 09:00:48 crc kubenswrapper[4934]: I1227 09:00:48.488141 4934 scope.go:117] "RemoveContainer" containerID="67d26929fe9646ad72855b9fddb3d99b10353382c7b8a204f41ed68b92f2f8c7" Dec 27 09:00:49 crc kubenswrapper[4934]: I1227 09:00:49.482919 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4080093-7ee5-483e-9422-ec3837273b04" path="/var/lib/kubelet/pods/d4080093-7ee5-483e-9422-ec3837273b04/volumes" Dec 27 09:01:00 crc kubenswrapper[4934]: I1227 09:01:00.184601 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29447101-6tmzs"] Dec 27 09:01:00 crc kubenswrapper[4934]: E1227 09:01:00.186397 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4080093-7ee5-483e-9422-ec3837273b04" containerName="registry-server" Dec 27 09:01:00 crc kubenswrapper[4934]: I1227 09:01:00.186448 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4080093-7ee5-483e-9422-ec3837273b04" containerName="registry-server" Dec 27 09:01:00 crc kubenswrapper[4934]: E1227 09:01:00.186507 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4080093-7ee5-483e-9422-ec3837273b04" containerName="extract-utilities" Dec 27 09:01:00 crc kubenswrapper[4934]: I1227 09:01:00.186527 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4080093-7ee5-483e-9422-ec3837273b04" containerName="extract-utilities" Dec 27 09:01:00 crc kubenswrapper[4934]: E1227 09:01:00.186591 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4080093-7ee5-483e-9422-ec3837273b04" containerName="extract-content" Dec 27 09:01:00 crc kubenswrapper[4934]: I1227 09:01:00.186615 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4080093-7ee5-483e-9422-ec3837273b04" containerName="extract-content" Dec 27 09:01:00 crc kubenswrapper[4934]: I1227 09:01:00.187306 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4080093-7ee5-483e-9422-ec3837273b04" containerName="registry-server" Dec 27 09:01:00 crc kubenswrapper[4934]: I1227 09:01:00.189409 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29447101-6tmzs" Dec 27 09:01:00 crc kubenswrapper[4934]: I1227 09:01:00.211403 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29447101-6tmzs"] Dec 27 09:01:00 crc kubenswrapper[4934]: I1227 09:01:00.258964 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/81559011-a562-4a12-9f3b-2c6286586406-fernet-keys\") pod \"keystone-cron-29447101-6tmzs\" (UID: \"81559011-a562-4a12-9f3b-2c6286586406\") " pod="openstack/keystone-cron-29447101-6tmzs" Dec 27 09:01:00 crc kubenswrapper[4934]: I1227 09:01:00.259055 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81559011-a562-4a12-9f3b-2c6286586406-combined-ca-bundle\") pod \"keystone-cron-29447101-6tmzs\" (UID: \"81559011-a562-4a12-9f3b-2c6286586406\") " pod="openstack/keystone-cron-29447101-6tmzs" Dec 27 09:01:00 crc kubenswrapper[4934]: I1227 09:01:00.259158 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqnc7\" (UniqueName: \"kubernetes.io/projected/81559011-a562-4a12-9f3b-2c6286586406-kube-api-access-zqnc7\") pod \"keystone-cron-29447101-6tmzs\" (UID: \"81559011-a562-4a12-9f3b-2c6286586406\") " pod="openstack/keystone-cron-29447101-6tmzs" Dec 27 09:01:00 crc kubenswrapper[4934]: I1227 09:01:00.259201 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81559011-a562-4a12-9f3b-2c6286586406-config-data\") pod \"keystone-cron-29447101-6tmzs\" (UID: \"81559011-a562-4a12-9f3b-2c6286586406\") " pod="openstack/keystone-cron-29447101-6tmzs" Dec 27 09:01:00 crc kubenswrapper[4934]: I1227 09:01:00.360883 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81559011-a562-4a12-9f3b-2c6286586406-combined-ca-bundle\") pod \"keystone-cron-29447101-6tmzs\" (UID: \"81559011-a562-4a12-9f3b-2c6286586406\") " pod="openstack/keystone-cron-29447101-6tmzs" Dec 27 09:01:00 crc kubenswrapper[4934]: I1227 09:01:00.360942 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqnc7\" (UniqueName: \"kubernetes.io/projected/81559011-a562-4a12-9f3b-2c6286586406-kube-api-access-zqnc7\") pod \"keystone-cron-29447101-6tmzs\" (UID: \"81559011-a562-4a12-9f3b-2c6286586406\") " pod="openstack/keystone-cron-29447101-6tmzs" Dec 27 09:01:00 crc kubenswrapper[4934]: I1227 09:01:00.360992 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81559011-a562-4a12-9f3b-2c6286586406-config-data\") pod \"keystone-cron-29447101-6tmzs\" (UID: \"81559011-a562-4a12-9f3b-2c6286586406\") " pod="openstack/keystone-cron-29447101-6tmzs" Dec 27 09:01:00 crc kubenswrapper[4934]: I1227 09:01:00.361202 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/81559011-a562-4a12-9f3b-2c6286586406-fernet-keys\") pod \"keystone-cron-29447101-6tmzs\" (UID: \"81559011-a562-4a12-9f3b-2c6286586406\") " pod="openstack/keystone-cron-29447101-6tmzs" Dec 27 09:01:00 crc kubenswrapper[4934]: I1227 09:01:00.741215 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/81559011-a562-4a12-9f3b-2c6286586406-fernet-keys\") pod \"keystone-cron-29447101-6tmzs\" (UID: \"81559011-a562-4a12-9f3b-2c6286586406\") " pod="openstack/keystone-cron-29447101-6tmzs" Dec 27 09:01:00 crc kubenswrapper[4934]: I1227 09:01:00.742223 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81559011-a562-4a12-9f3b-2c6286586406-config-data\") pod \"keystone-cron-29447101-6tmzs\" (UID: \"81559011-a562-4a12-9f3b-2c6286586406\") " pod="openstack/keystone-cron-29447101-6tmzs" Dec 27 09:01:00 crc kubenswrapper[4934]: I1227 09:01:00.742435 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81559011-a562-4a12-9f3b-2c6286586406-combined-ca-bundle\") pod \"keystone-cron-29447101-6tmzs\" (UID: \"81559011-a562-4a12-9f3b-2c6286586406\") " pod="openstack/keystone-cron-29447101-6tmzs" Dec 27 09:01:00 crc kubenswrapper[4934]: I1227 09:01:00.743011 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqnc7\" (UniqueName: \"kubernetes.io/projected/81559011-a562-4a12-9f3b-2c6286586406-kube-api-access-zqnc7\") pod \"keystone-cron-29447101-6tmzs\" (UID: \"81559011-a562-4a12-9f3b-2c6286586406\") " pod="openstack/keystone-cron-29447101-6tmzs" Dec 27 09:01:00 crc kubenswrapper[4934]: I1227 09:01:00.827399 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29447101-6tmzs" Dec 27 09:01:01 crc kubenswrapper[4934]: I1227 09:01:01.374824 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29447101-6tmzs"] Dec 27 09:01:01 crc kubenswrapper[4934]: W1227 09:01:01.381945 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81559011_a562_4a12_9f3b_2c6286586406.slice/crio-b879007019329979b9b506c5d8af67f0c128c6d46d5cb7d6d53289ba6d73ae4a WatchSource:0}: Error finding container b879007019329979b9b506c5d8af67f0c128c6d46d5cb7d6d53289ba6d73ae4a: Status 404 returned error can't find the container with id b879007019329979b9b506c5d8af67f0c128c6d46d5cb7d6d53289ba6d73ae4a Dec 27 09:01:01 crc kubenswrapper[4934]: I1227 09:01:01.567397 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29447101-6tmzs" event={"ID":"81559011-a562-4a12-9f3b-2c6286586406","Type":"ContainerStarted","Data":"b879007019329979b9b506c5d8af67f0c128c6d46d5cb7d6d53289ba6d73ae4a"} Dec 27 09:01:02 crc kubenswrapper[4934]: I1227 09:01:02.588493 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29447101-6tmzs" event={"ID":"81559011-a562-4a12-9f3b-2c6286586406","Type":"ContainerStarted","Data":"e39123d1f95f499455a3b2d851ffa33c743050a0007f6810f1f2d0ad30fc5439"} Dec 27 09:01:02 crc kubenswrapper[4934]: I1227 09:01:02.603566 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29447101-6tmzs" podStartSLOduration=2.603544108 podStartE2EDuration="2.603544108s" podCreationTimestamp="2025-12-27 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 09:01:02.603137798 +0000 UTC m=+4723.423578412" watchObservedRunningTime="2025-12-27 09:01:02.603544108 +0000 UTC m=+4723.423984722" Dec 27 09:01:05 crc kubenswrapper[4934]: I1227 09:01:05.624736 4934 generic.go:334] "Generic (PLEG): container finished" podID="81559011-a562-4a12-9f3b-2c6286586406" containerID="e39123d1f95f499455a3b2d851ffa33c743050a0007f6810f1f2d0ad30fc5439" exitCode=0 Dec 27 09:01:05 crc kubenswrapper[4934]: I1227 09:01:05.625408 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29447101-6tmzs" event={"ID":"81559011-a562-4a12-9f3b-2c6286586406","Type":"ContainerDied","Data":"e39123d1f95f499455a3b2d851ffa33c743050a0007f6810f1f2d0ad30fc5439"} Dec 27 09:01:07 crc kubenswrapper[4934]: I1227 09:01:07.084037 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29447101-6tmzs" Dec 27 09:01:07 crc kubenswrapper[4934]: I1227 09:01:07.242714 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/81559011-a562-4a12-9f3b-2c6286586406-fernet-keys\") pod \"81559011-a562-4a12-9f3b-2c6286586406\" (UID: \"81559011-a562-4a12-9f3b-2c6286586406\") " Dec 27 09:01:07 crc kubenswrapper[4934]: I1227 09:01:07.242962 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqnc7\" (UniqueName: \"kubernetes.io/projected/81559011-a562-4a12-9f3b-2c6286586406-kube-api-access-zqnc7\") pod \"81559011-a562-4a12-9f3b-2c6286586406\" (UID: \"81559011-a562-4a12-9f3b-2c6286586406\") " Dec 27 09:01:07 crc kubenswrapper[4934]: I1227 09:01:07.243331 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81559011-a562-4a12-9f3b-2c6286586406-combined-ca-bundle\") pod \"81559011-a562-4a12-9f3b-2c6286586406\" (UID: \"81559011-a562-4a12-9f3b-2c6286586406\") " Dec 27 09:01:07 crc kubenswrapper[4934]: I1227 09:01:07.243400 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81559011-a562-4a12-9f3b-2c6286586406-config-data\") pod \"81559011-a562-4a12-9f3b-2c6286586406\" (UID: \"81559011-a562-4a12-9f3b-2c6286586406\") " Dec 27 09:01:07 crc kubenswrapper[4934]: I1227 09:01:07.251136 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81559011-a562-4a12-9f3b-2c6286586406-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "81559011-a562-4a12-9f3b-2c6286586406" (UID: "81559011-a562-4a12-9f3b-2c6286586406"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 09:01:07 crc kubenswrapper[4934]: I1227 09:01:07.252197 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81559011-a562-4a12-9f3b-2c6286586406-kube-api-access-zqnc7" (OuterVolumeSpecName: "kube-api-access-zqnc7") pod "81559011-a562-4a12-9f3b-2c6286586406" (UID: "81559011-a562-4a12-9f3b-2c6286586406"). InnerVolumeSpecName "kube-api-access-zqnc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:01:07 crc kubenswrapper[4934]: I1227 09:01:07.298767 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81559011-a562-4a12-9f3b-2c6286586406-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81559011-a562-4a12-9f3b-2c6286586406" (UID: "81559011-a562-4a12-9f3b-2c6286586406"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 09:01:07 crc kubenswrapper[4934]: I1227 09:01:07.343864 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81559011-a562-4a12-9f3b-2c6286586406-config-data" (OuterVolumeSpecName: "config-data") pod "81559011-a562-4a12-9f3b-2c6286586406" (UID: "81559011-a562-4a12-9f3b-2c6286586406"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 09:01:07 crc kubenswrapper[4934]: I1227 09:01:07.347295 4934 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/81559011-a562-4a12-9f3b-2c6286586406-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 27 09:01:07 crc kubenswrapper[4934]: I1227 09:01:07.347321 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqnc7\" (UniqueName: \"kubernetes.io/projected/81559011-a562-4a12-9f3b-2c6286586406-kube-api-access-zqnc7\") on node \"crc\" DevicePath \"\"" Dec 27 09:01:07 crc kubenswrapper[4934]: I1227 09:01:07.347333 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81559011-a562-4a12-9f3b-2c6286586406-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 09:01:07 crc kubenswrapper[4934]: I1227 09:01:07.347345 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81559011-a562-4a12-9f3b-2c6286586406-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 09:01:07 crc kubenswrapper[4934]: I1227 09:01:07.663154 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29447101-6tmzs" event={"ID":"81559011-a562-4a12-9f3b-2c6286586406","Type":"ContainerDied","Data":"b879007019329979b9b506c5d8af67f0c128c6d46d5cb7d6d53289ba6d73ae4a"} Dec 27 09:01:07 crc kubenswrapper[4934]: I1227 09:01:07.663492 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b879007019329979b9b506c5d8af67f0c128c6d46d5cb7d6d53289ba6d73ae4a" Dec 27 09:01:07 crc kubenswrapper[4934]: I1227 09:01:07.663585 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29447101-6tmzs" Dec 27 09:01:40 crc kubenswrapper[4934]: I1227 09:01:40.462361 4934 trace.go:236] Trace[515292657]: "Calculate volume metrics of persistence for pod openstack/rabbitmq-server-1" (27-Dec-2025 09:01:39.318) (total time: 1143ms): Dec 27 09:01:40 crc kubenswrapper[4934]: Trace[515292657]: [1.143834268s] [1.143834268s] END Dec 27 09:02:15 crc kubenswrapper[4934]: I1227 09:02:15.329771 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 09:02:15 crc kubenswrapper[4934]: I1227 09:02:15.330465 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 09:02:43 crc kubenswrapper[4934]: I1227 09:02:43.320322 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-57s9g"] Dec 27 09:02:43 crc kubenswrapper[4934]: E1227 09:02:43.322488 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81559011-a562-4a12-9f3b-2c6286586406" containerName="keystone-cron" Dec 27 09:02:43 crc kubenswrapper[4934]: I1227 09:02:43.322607 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="81559011-a562-4a12-9f3b-2c6286586406" containerName="keystone-cron" Dec 27 09:02:43 crc kubenswrapper[4934]: I1227 09:02:43.323000 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="81559011-a562-4a12-9f3b-2c6286586406" containerName="keystone-cron" Dec 27 09:02:43 crc kubenswrapper[4934]: I1227 09:02:43.325237 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-57s9g" Dec 27 09:02:43 crc kubenswrapper[4934]: I1227 09:02:43.347881 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-57s9g"] Dec 27 09:02:43 crc kubenswrapper[4934]: I1227 09:02:43.413947 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj4ft\" (UniqueName: \"kubernetes.io/projected/97a6295d-e613-43f2-a1ae-4ae80e2b42f7-kube-api-access-zj4ft\") pod \"redhat-operators-57s9g\" (UID: \"97a6295d-e613-43f2-a1ae-4ae80e2b42f7\") " pod="openshift-marketplace/redhat-operators-57s9g" Dec 27 09:02:43 crc kubenswrapper[4934]: I1227 09:02:43.414071 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97a6295d-e613-43f2-a1ae-4ae80e2b42f7-catalog-content\") pod \"redhat-operators-57s9g\" (UID: \"97a6295d-e613-43f2-a1ae-4ae80e2b42f7\") " pod="openshift-marketplace/redhat-operators-57s9g" Dec 27 09:02:43 crc kubenswrapper[4934]: I1227 09:02:43.414170 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97a6295d-e613-43f2-a1ae-4ae80e2b42f7-utilities\") pod \"redhat-operators-57s9g\" (UID: \"97a6295d-e613-43f2-a1ae-4ae80e2b42f7\") " pod="openshift-marketplace/redhat-operators-57s9g" Dec 27 09:02:43 crc kubenswrapper[4934]: I1227 09:02:43.517632 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97a6295d-e613-43f2-a1ae-4ae80e2b42f7-catalog-content\") pod \"redhat-operators-57s9g\" (UID: \"97a6295d-e613-43f2-a1ae-4ae80e2b42f7\") " pod="openshift-marketplace/redhat-operators-57s9g" Dec 27 09:02:43 crc kubenswrapper[4934]: I1227 09:02:43.517939 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97a6295d-e613-43f2-a1ae-4ae80e2b42f7-catalog-content\") pod \"redhat-operators-57s9g\" (UID: \"97a6295d-e613-43f2-a1ae-4ae80e2b42f7\") " pod="openshift-marketplace/redhat-operators-57s9g" Dec 27 09:02:43 crc kubenswrapper[4934]: I1227 09:02:43.518705 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97a6295d-e613-43f2-a1ae-4ae80e2b42f7-utilities\") pod \"redhat-operators-57s9g\" (UID: \"97a6295d-e613-43f2-a1ae-4ae80e2b42f7\") " pod="openshift-marketplace/redhat-operators-57s9g" Dec 27 09:02:43 crc kubenswrapper[4934]: I1227 09:02:43.518972 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj4ft\" (UniqueName: \"kubernetes.io/projected/97a6295d-e613-43f2-a1ae-4ae80e2b42f7-kube-api-access-zj4ft\") pod \"redhat-operators-57s9g\" (UID: \"97a6295d-e613-43f2-a1ae-4ae80e2b42f7\") " pod="openshift-marketplace/redhat-operators-57s9g" Dec 27 09:02:43 crc kubenswrapper[4934]: I1227 09:02:43.519325 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97a6295d-e613-43f2-a1ae-4ae80e2b42f7-utilities\") pod \"redhat-operators-57s9g\" (UID: \"97a6295d-e613-43f2-a1ae-4ae80e2b42f7\") " pod="openshift-marketplace/redhat-operators-57s9g" Dec 27 09:02:43 crc kubenswrapper[4934]: I1227 09:02:43.541296 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj4ft\" (UniqueName: \"kubernetes.io/projected/97a6295d-e613-43f2-a1ae-4ae80e2b42f7-kube-api-access-zj4ft\") pod \"redhat-operators-57s9g\" (UID: \"97a6295d-e613-43f2-a1ae-4ae80e2b42f7\") " pod="openshift-marketplace/redhat-operators-57s9g" Dec 27 09:02:43 crc kubenswrapper[4934]: I1227 09:02:43.654210 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-57s9g" Dec 27 09:02:44 crc kubenswrapper[4934]: I1227 09:02:44.172186 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-57s9g"] Dec 27 09:02:44 crc kubenswrapper[4934]: I1227 09:02:44.865545 4934 generic.go:334] "Generic (PLEG): container finished" podID="97a6295d-e613-43f2-a1ae-4ae80e2b42f7" containerID="72dac9d4b98af4f16f1e27fbb67833531667e288fccf7785157ccab6c2e8b987" exitCode=0 Dec 27 09:02:44 crc kubenswrapper[4934]: I1227 09:02:44.865864 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-57s9g" event={"ID":"97a6295d-e613-43f2-a1ae-4ae80e2b42f7","Type":"ContainerDied","Data":"72dac9d4b98af4f16f1e27fbb67833531667e288fccf7785157ccab6c2e8b987"} Dec 27 09:02:44 crc kubenswrapper[4934]: I1227 09:02:44.865895 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-57s9g" event={"ID":"97a6295d-e613-43f2-a1ae-4ae80e2b42f7","Type":"ContainerStarted","Data":"f1a6151cd4fc129fb96b02a5c33a9afc7aad09e39b75df088b439470c43a4be4"} Dec 27 09:02:44 crc kubenswrapper[4934]: I1227 09:02:44.867854 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 27 09:02:45 crc kubenswrapper[4934]: I1227 09:02:45.330648 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 09:02:45 crc kubenswrapper[4934]: I1227 09:02:45.331116 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 09:02:45 crc kubenswrapper[4934]: I1227 09:02:45.884893 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-57s9g" event={"ID":"97a6295d-e613-43f2-a1ae-4ae80e2b42f7","Type":"ContainerStarted","Data":"cae8a9da836a94620299773fbfef723623ca22bb429332c322c4d895bfb43533"} Dec 27 09:02:50 crc kubenswrapper[4934]: I1227 09:02:50.947210 4934 generic.go:334] "Generic (PLEG): container finished" podID="97a6295d-e613-43f2-a1ae-4ae80e2b42f7" containerID="cae8a9da836a94620299773fbfef723623ca22bb429332c322c4d895bfb43533" exitCode=0 Dec 27 09:02:50 crc kubenswrapper[4934]: I1227 09:02:50.947303 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-57s9g" event={"ID":"97a6295d-e613-43f2-a1ae-4ae80e2b42f7","Type":"ContainerDied","Data":"cae8a9da836a94620299773fbfef723623ca22bb429332c322c4d895bfb43533"} Dec 27 09:02:51 crc kubenswrapper[4934]: I1227 09:02:51.977161 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-57s9g" event={"ID":"97a6295d-e613-43f2-a1ae-4ae80e2b42f7","Type":"ContainerStarted","Data":"2fe419065b547701ccec2482471f6a79002329ff9b93f456170003aa9f1fe01a"} Dec 27 09:02:52 crc kubenswrapper[4934]: I1227 09:02:52.005118 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-57s9g" podStartSLOduration=2.540472276 podStartE2EDuration="9.005100159s" podCreationTimestamp="2025-12-27 09:02:43 +0000 UTC" firstStartedPulling="2025-12-27 09:02:44.867615725 +0000 UTC m=+4825.688056319" lastFinishedPulling="2025-12-27 09:02:51.332243608 +0000 UTC m=+4832.152684202" observedRunningTime="2025-12-27 09:02:52.00031521 +0000 UTC m=+4832.820755814" watchObservedRunningTime="2025-12-27 09:02:52.005100159 +0000 UTC m=+4832.825540743" Dec 27 09:02:53 crc kubenswrapper[4934]: I1227 09:02:53.654788 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-57s9g" Dec 27 09:02:53 crc kubenswrapper[4934]: I1227 09:02:53.655126 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-57s9g" Dec 27 09:02:54 crc kubenswrapper[4934]: I1227 09:02:54.707128 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-57s9g" podUID="97a6295d-e613-43f2-a1ae-4ae80e2b42f7" containerName="registry-server" probeResult="failure" output=< Dec 27 09:02:54 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 09:02:54 crc kubenswrapper[4934]: > Dec 27 09:02:58 crc kubenswrapper[4934]: I1227 09:02:58.494760 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xglll"] Dec 27 09:02:58 crc kubenswrapper[4934]: I1227 09:02:58.498433 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xglll" Dec 27 09:02:58 crc kubenswrapper[4934]: I1227 09:02:58.509800 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xglll"] Dec 27 09:02:58 crc kubenswrapper[4934]: I1227 09:02:58.552566 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgnj4\" (UniqueName: \"kubernetes.io/projected/175a7c54-e0a9-4f22-8d6d-f74422c2505c-kube-api-access-dgnj4\") pod \"redhat-marketplace-xglll\" (UID: \"175a7c54-e0a9-4f22-8d6d-f74422c2505c\") " pod="openshift-marketplace/redhat-marketplace-xglll" Dec 27 09:02:58 crc kubenswrapper[4934]: I1227 09:02:58.552913 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/175a7c54-e0a9-4f22-8d6d-f74422c2505c-catalog-content\") pod \"redhat-marketplace-xglll\" (UID: \"175a7c54-e0a9-4f22-8d6d-f74422c2505c\") " pod="openshift-marketplace/redhat-marketplace-xglll" Dec 27 09:02:58 crc kubenswrapper[4934]: I1227 09:02:58.552946 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/175a7c54-e0a9-4f22-8d6d-f74422c2505c-utilities\") pod \"redhat-marketplace-xglll\" (UID: \"175a7c54-e0a9-4f22-8d6d-f74422c2505c\") " pod="openshift-marketplace/redhat-marketplace-xglll" Dec 27 09:02:58 crc kubenswrapper[4934]: I1227 09:02:58.655153 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/175a7c54-e0a9-4f22-8d6d-f74422c2505c-catalog-content\") pod \"redhat-marketplace-xglll\" (UID: \"175a7c54-e0a9-4f22-8d6d-f74422c2505c\") " pod="openshift-marketplace/redhat-marketplace-xglll" Dec 27 09:02:58 crc kubenswrapper[4934]: I1227 09:02:58.655210 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/175a7c54-e0a9-4f22-8d6d-f74422c2505c-utilities\") pod \"redhat-marketplace-xglll\" (UID: \"175a7c54-e0a9-4f22-8d6d-f74422c2505c\") " pod="openshift-marketplace/redhat-marketplace-xglll" Dec 27 09:02:58 crc kubenswrapper[4934]: I1227 09:02:58.655419 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgnj4\" (UniqueName: \"kubernetes.io/projected/175a7c54-e0a9-4f22-8d6d-f74422c2505c-kube-api-access-dgnj4\") pod \"redhat-marketplace-xglll\" (UID: \"175a7c54-e0a9-4f22-8d6d-f74422c2505c\") " pod="openshift-marketplace/redhat-marketplace-xglll" Dec 27 09:02:58 crc kubenswrapper[4934]: I1227 09:02:58.655878 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/175a7c54-e0a9-4f22-8d6d-f74422c2505c-catalog-content\") pod \"redhat-marketplace-xglll\" (UID: \"175a7c54-e0a9-4f22-8d6d-f74422c2505c\") " pod="openshift-marketplace/redhat-marketplace-xglll" Dec 27 09:02:58 crc kubenswrapper[4934]: I1227 09:02:58.655878 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/175a7c54-e0a9-4f22-8d6d-f74422c2505c-utilities\") pod \"redhat-marketplace-xglll\" (UID: \"175a7c54-e0a9-4f22-8d6d-f74422c2505c\") " pod="openshift-marketplace/redhat-marketplace-xglll" Dec 27 09:02:58 crc kubenswrapper[4934]: I1227 09:02:58.741189 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgnj4\" (UniqueName: \"kubernetes.io/projected/175a7c54-e0a9-4f22-8d6d-f74422c2505c-kube-api-access-dgnj4\") pod \"redhat-marketplace-xglll\" (UID: \"175a7c54-e0a9-4f22-8d6d-f74422c2505c\") " pod="openshift-marketplace/redhat-marketplace-xglll" Dec 27 09:02:58 crc kubenswrapper[4934]: I1227 09:02:58.848274 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xglll" Dec 27 09:02:59 crc kubenswrapper[4934]: W1227 09:02:59.412277 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod175a7c54_e0a9_4f22_8d6d_f74422c2505c.slice/crio-f67d1db50c5f9fc906ab9d173449ffcb49dee96968feded4d0a307f31f4276e3 WatchSource:0}: Error finding container f67d1db50c5f9fc906ab9d173449ffcb49dee96968feded4d0a307f31f4276e3: Status 404 returned error can't find the container with id f67d1db50c5f9fc906ab9d173449ffcb49dee96968feded4d0a307f31f4276e3 Dec 27 09:02:59 crc kubenswrapper[4934]: I1227 09:02:59.415220 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xglll"] Dec 27 09:03:00 crc kubenswrapper[4934]: I1227 09:03:00.078313 4934 generic.go:334] "Generic (PLEG): container finished" podID="175a7c54-e0a9-4f22-8d6d-f74422c2505c" containerID="c7e79a36cc2514415e6dd0ef7bab82e6c91a72928d54da9b6333aece8431404c" exitCode=0 Dec 27 09:03:00 crc kubenswrapper[4934]: I1227 09:03:00.078381 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xglll" event={"ID":"175a7c54-e0a9-4f22-8d6d-f74422c2505c","Type":"ContainerDied","Data":"c7e79a36cc2514415e6dd0ef7bab82e6c91a72928d54da9b6333aece8431404c"} Dec 27 09:03:00 crc kubenswrapper[4934]: I1227 09:03:00.079999 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xglll" event={"ID":"175a7c54-e0a9-4f22-8d6d-f74422c2505c","Type":"ContainerStarted","Data":"f67d1db50c5f9fc906ab9d173449ffcb49dee96968feded4d0a307f31f4276e3"} Dec 27 09:03:02 crc kubenswrapper[4934]: I1227 09:03:02.104112 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xglll" event={"ID":"175a7c54-e0a9-4f22-8d6d-f74422c2505c","Type":"ContainerStarted","Data":"ca47097cf575b785676ef8415ed49fca93c5e5b10befd265b53de93df58393c8"} Dec 27 09:03:03 crc kubenswrapper[4934]: I1227 09:03:03.121895 4934 generic.go:334] "Generic (PLEG): container finished" podID="175a7c54-e0a9-4f22-8d6d-f74422c2505c" containerID="ca47097cf575b785676ef8415ed49fca93c5e5b10befd265b53de93df58393c8" exitCode=0 Dec 27 09:03:03 crc kubenswrapper[4934]: I1227 09:03:03.122212 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xglll" event={"ID":"175a7c54-e0a9-4f22-8d6d-f74422c2505c","Type":"ContainerDied","Data":"ca47097cf575b785676ef8415ed49fca93c5e5b10befd265b53de93df58393c8"} Dec 27 09:03:04 crc kubenswrapper[4934]: I1227 09:03:04.135712 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xglll" event={"ID":"175a7c54-e0a9-4f22-8d6d-f74422c2505c","Type":"ContainerStarted","Data":"c7df9072a0643a0a469745be5cf507b5954793887da6d1ae6079354c3534b198"} Dec 27 09:03:04 crc kubenswrapper[4934]: I1227 09:03:04.163867 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xglll" podStartSLOduration=2.449145342 podStartE2EDuration="6.163840268s" podCreationTimestamp="2025-12-27 09:02:58 +0000 UTC" firstStartedPulling="2025-12-27 09:03:00.081643484 +0000 UTC m=+4840.902084078" lastFinishedPulling="2025-12-27 09:03:03.79633841 +0000 UTC m=+4844.616779004" observedRunningTime="2025-12-27 09:03:04.153893811 +0000 UTC m=+4844.974334425" watchObservedRunningTime="2025-12-27 09:03:04.163840268 +0000 UTC m=+4844.984280862" Dec 27 09:03:04 crc kubenswrapper[4934]: I1227 09:03:04.716433 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-57s9g" podUID="97a6295d-e613-43f2-a1ae-4ae80e2b42f7" containerName="registry-server" probeResult="failure" output=< Dec 27 09:03:04 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 09:03:04 crc kubenswrapper[4934]: > Dec 27 09:03:08 crc kubenswrapper[4934]: I1227 09:03:08.848989 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xglll" Dec 27 09:03:08 crc kubenswrapper[4934]: I1227 09:03:08.849458 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xglll" Dec 27 09:03:08 crc kubenswrapper[4934]: I1227 09:03:08.896420 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xglll" Dec 27 09:03:09 crc kubenswrapper[4934]: I1227 09:03:09.251772 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xglll" Dec 27 09:03:09 crc kubenswrapper[4934]: I1227 09:03:09.320378 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xglll"] Dec 27 09:03:11 crc kubenswrapper[4934]: I1227 09:03:11.219656 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xglll" podUID="175a7c54-e0a9-4f22-8d6d-f74422c2505c" containerName="registry-server" containerID="cri-o://c7df9072a0643a0a469745be5cf507b5954793887da6d1ae6079354c3534b198" gracePeriod=2 Dec 27 09:03:11 crc kubenswrapper[4934]: I1227 09:03:11.760030 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xglll" Dec 27 09:03:11 crc kubenswrapper[4934]: I1227 09:03:11.885320 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgnj4\" (UniqueName: \"kubernetes.io/projected/175a7c54-e0a9-4f22-8d6d-f74422c2505c-kube-api-access-dgnj4\") pod \"175a7c54-e0a9-4f22-8d6d-f74422c2505c\" (UID: \"175a7c54-e0a9-4f22-8d6d-f74422c2505c\") " Dec 27 09:03:11 crc kubenswrapper[4934]: I1227 09:03:11.885803 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/175a7c54-e0a9-4f22-8d6d-f74422c2505c-utilities\") pod \"175a7c54-e0a9-4f22-8d6d-f74422c2505c\" (UID: \"175a7c54-e0a9-4f22-8d6d-f74422c2505c\") " Dec 27 09:03:11 crc kubenswrapper[4934]: I1227 09:03:11.886193 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/175a7c54-e0a9-4f22-8d6d-f74422c2505c-catalog-content\") pod \"175a7c54-e0a9-4f22-8d6d-f74422c2505c\" (UID: \"175a7c54-e0a9-4f22-8d6d-f74422c2505c\") " Dec 27 09:03:11 crc kubenswrapper[4934]: I1227 09:03:11.887074 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/175a7c54-e0a9-4f22-8d6d-f74422c2505c-utilities" (OuterVolumeSpecName: "utilities") pod "175a7c54-e0a9-4f22-8d6d-f74422c2505c" (UID: "175a7c54-e0a9-4f22-8d6d-f74422c2505c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:03:11 crc kubenswrapper[4934]: I1227 09:03:11.899572 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/175a7c54-e0a9-4f22-8d6d-f74422c2505c-kube-api-access-dgnj4" (OuterVolumeSpecName: "kube-api-access-dgnj4") pod "175a7c54-e0a9-4f22-8d6d-f74422c2505c" (UID: "175a7c54-e0a9-4f22-8d6d-f74422c2505c"). InnerVolumeSpecName "kube-api-access-dgnj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:03:11 crc kubenswrapper[4934]: I1227 09:03:11.918051 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/175a7c54-e0a9-4f22-8d6d-f74422c2505c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "175a7c54-e0a9-4f22-8d6d-f74422c2505c" (UID: "175a7c54-e0a9-4f22-8d6d-f74422c2505c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:03:11 crc kubenswrapper[4934]: I1227 09:03:11.989416 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/175a7c54-e0a9-4f22-8d6d-f74422c2505c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 09:03:11 crc kubenswrapper[4934]: I1227 09:03:11.989458 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgnj4\" (UniqueName: \"kubernetes.io/projected/175a7c54-e0a9-4f22-8d6d-f74422c2505c-kube-api-access-dgnj4\") on node \"crc\" DevicePath \"\"" Dec 27 09:03:11 crc kubenswrapper[4934]: I1227 09:03:11.989472 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/175a7c54-e0a9-4f22-8d6d-f74422c2505c-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 09:03:12 crc kubenswrapper[4934]: I1227 09:03:12.230711 4934 generic.go:334] "Generic (PLEG): container finished" podID="175a7c54-e0a9-4f22-8d6d-f74422c2505c" containerID="c7df9072a0643a0a469745be5cf507b5954793887da6d1ae6079354c3534b198" exitCode=0 Dec 27 09:03:12 crc kubenswrapper[4934]: I1227 09:03:12.230750 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xglll" event={"ID":"175a7c54-e0a9-4f22-8d6d-f74422c2505c","Type":"ContainerDied","Data":"c7df9072a0643a0a469745be5cf507b5954793887da6d1ae6079354c3534b198"} Dec 27 09:03:12 crc kubenswrapper[4934]: I1227 09:03:12.230774 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xglll" event={"ID":"175a7c54-e0a9-4f22-8d6d-f74422c2505c","Type":"ContainerDied","Data":"f67d1db50c5f9fc906ab9d173449ffcb49dee96968feded4d0a307f31f4276e3"} Dec 27 09:03:12 crc kubenswrapper[4934]: I1227 09:03:12.230795 4934 scope.go:117] "RemoveContainer" containerID="c7df9072a0643a0a469745be5cf507b5954793887da6d1ae6079354c3534b198" Dec 27 09:03:12 crc kubenswrapper[4934]: I1227 09:03:12.230904 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xglll" Dec 27 09:03:12 crc kubenswrapper[4934]: I1227 09:03:12.267407 4934 scope.go:117] "RemoveContainer" containerID="ca47097cf575b785676ef8415ed49fca93c5e5b10befd265b53de93df58393c8" Dec 27 09:03:12 crc kubenswrapper[4934]: I1227 09:03:12.279137 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xglll"] Dec 27 09:03:12 crc kubenswrapper[4934]: I1227 09:03:12.323374 4934 scope.go:117] "RemoveContainer" containerID="c7e79a36cc2514415e6dd0ef7bab82e6c91a72928d54da9b6333aece8431404c" Dec 27 09:03:12 crc kubenswrapper[4934]: I1227 09:03:12.323617 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xglll"] Dec 27 09:03:12 crc kubenswrapper[4934]: I1227 09:03:12.363820 4934 scope.go:117] "RemoveContainer" containerID="c7df9072a0643a0a469745be5cf507b5954793887da6d1ae6079354c3534b198" Dec 27 09:03:12 crc kubenswrapper[4934]: E1227 09:03:12.364392 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7df9072a0643a0a469745be5cf507b5954793887da6d1ae6079354c3534b198\": container with ID starting with c7df9072a0643a0a469745be5cf507b5954793887da6d1ae6079354c3534b198 not found: ID does not exist" containerID="c7df9072a0643a0a469745be5cf507b5954793887da6d1ae6079354c3534b198" Dec 27 09:03:12 crc kubenswrapper[4934]: I1227 09:03:12.364459 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7df9072a0643a0a469745be5cf507b5954793887da6d1ae6079354c3534b198"} err="failed to get container status \"c7df9072a0643a0a469745be5cf507b5954793887da6d1ae6079354c3534b198\": rpc error: code = NotFound desc = could not find container \"c7df9072a0643a0a469745be5cf507b5954793887da6d1ae6079354c3534b198\": container with ID starting with c7df9072a0643a0a469745be5cf507b5954793887da6d1ae6079354c3534b198 not found: ID does not exist" Dec 27 09:03:12 crc kubenswrapper[4934]: I1227 09:03:12.364493 4934 scope.go:117] "RemoveContainer" containerID="ca47097cf575b785676ef8415ed49fca93c5e5b10befd265b53de93df58393c8" Dec 27 09:03:12 crc kubenswrapper[4934]: E1227 09:03:12.364901 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca47097cf575b785676ef8415ed49fca93c5e5b10befd265b53de93df58393c8\": container with ID starting with ca47097cf575b785676ef8415ed49fca93c5e5b10befd265b53de93df58393c8 not found: ID does not exist" containerID="ca47097cf575b785676ef8415ed49fca93c5e5b10befd265b53de93df58393c8" Dec 27 09:03:12 crc kubenswrapper[4934]: I1227 09:03:12.364926 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca47097cf575b785676ef8415ed49fca93c5e5b10befd265b53de93df58393c8"} err="failed to get container status \"ca47097cf575b785676ef8415ed49fca93c5e5b10befd265b53de93df58393c8\": rpc error: code = NotFound desc = could not find container \"ca47097cf575b785676ef8415ed49fca93c5e5b10befd265b53de93df58393c8\": container with ID starting with ca47097cf575b785676ef8415ed49fca93c5e5b10befd265b53de93df58393c8 not found: ID does not exist" Dec 27 09:03:12 crc kubenswrapper[4934]: I1227 09:03:12.364939 4934 scope.go:117] "RemoveContainer" containerID="c7e79a36cc2514415e6dd0ef7bab82e6c91a72928d54da9b6333aece8431404c" Dec 27 09:03:12 crc kubenswrapper[4934]: E1227 09:03:12.365259 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7e79a36cc2514415e6dd0ef7bab82e6c91a72928d54da9b6333aece8431404c\": container with ID starting with c7e79a36cc2514415e6dd0ef7bab82e6c91a72928d54da9b6333aece8431404c not found: ID does not exist" containerID="c7e79a36cc2514415e6dd0ef7bab82e6c91a72928d54da9b6333aece8431404c" Dec 27 09:03:12 crc kubenswrapper[4934]: I1227 09:03:12.365282 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7e79a36cc2514415e6dd0ef7bab82e6c91a72928d54da9b6333aece8431404c"} err="failed to get container status \"c7e79a36cc2514415e6dd0ef7bab82e6c91a72928d54da9b6333aece8431404c\": rpc error: code = NotFound desc = could not find container \"c7e79a36cc2514415e6dd0ef7bab82e6c91a72928d54da9b6333aece8431404c\": container with ID starting with c7e79a36cc2514415e6dd0ef7bab82e6c91a72928d54da9b6333aece8431404c not found: ID does not exist" Dec 27 09:03:13 crc kubenswrapper[4934]: I1227 09:03:13.486990 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="175a7c54-e0a9-4f22-8d6d-f74422c2505c" path="/var/lib/kubelet/pods/175a7c54-e0a9-4f22-8d6d-f74422c2505c/volumes" Dec 27 09:03:13 crc kubenswrapper[4934]: I1227 09:03:13.704787 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-57s9g" Dec 27 09:03:13 crc kubenswrapper[4934]: I1227 09:03:13.753454 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-57s9g" Dec 27 09:03:14 crc kubenswrapper[4934]: I1227 09:03:14.547289 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-57s9g"] Dec 27 09:03:15 crc kubenswrapper[4934]: I1227 09:03:15.269785 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-57s9g" podUID="97a6295d-e613-43f2-a1ae-4ae80e2b42f7" containerName="registry-server" containerID="cri-o://2fe419065b547701ccec2482471f6a79002329ff9b93f456170003aa9f1fe01a" gracePeriod=2 Dec 27 09:03:15 crc kubenswrapper[4934]: I1227 09:03:15.329658 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 09:03:15 crc kubenswrapper[4934]: I1227 09:03:15.329709 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 09:03:15 crc kubenswrapper[4934]: I1227 09:03:15.329762 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 09:03:15 crc kubenswrapper[4934]: I1227 09:03:15.330960 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8224ba78a25faf72ee4cb82888cba291e56d3f159515c14eec31cad4a6c0f282"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 09:03:15 crc kubenswrapper[4934]: I1227 09:03:15.331037 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://8224ba78a25faf72ee4cb82888cba291e56d3f159515c14eec31cad4a6c0f282" gracePeriod=600 Dec 27 09:03:15 crc kubenswrapper[4934]: I1227 09:03:15.860459 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-57s9g" Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.007868 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zj4ft\" (UniqueName: \"kubernetes.io/projected/97a6295d-e613-43f2-a1ae-4ae80e2b42f7-kube-api-access-zj4ft\") pod \"97a6295d-e613-43f2-a1ae-4ae80e2b42f7\" (UID: \"97a6295d-e613-43f2-a1ae-4ae80e2b42f7\") " Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.008152 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97a6295d-e613-43f2-a1ae-4ae80e2b42f7-catalog-content\") pod \"97a6295d-e613-43f2-a1ae-4ae80e2b42f7\" (UID: \"97a6295d-e613-43f2-a1ae-4ae80e2b42f7\") " Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.008362 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97a6295d-e613-43f2-a1ae-4ae80e2b42f7-utilities\") pod \"97a6295d-e613-43f2-a1ae-4ae80e2b42f7\" (UID: \"97a6295d-e613-43f2-a1ae-4ae80e2b42f7\") " Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.009383 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97a6295d-e613-43f2-a1ae-4ae80e2b42f7-utilities" (OuterVolumeSpecName: "utilities") pod "97a6295d-e613-43f2-a1ae-4ae80e2b42f7" (UID: "97a6295d-e613-43f2-a1ae-4ae80e2b42f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.010158 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97a6295d-e613-43f2-a1ae-4ae80e2b42f7-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.014193 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97a6295d-e613-43f2-a1ae-4ae80e2b42f7-kube-api-access-zj4ft" (OuterVolumeSpecName: "kube-api-access-zj4ft") pod "97a6295d-e613-43f2-a1ae-4ae80e2b42f7" (UID: "97a6295d-e613-43f2-a1ae-4ae80e2b42f7"). InnerVolumeSpecName "kube-api-access-zj4ft". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.112684 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zj4ft\" (UniqueName: \"kubernetes.io/projected/97a6295d-e613-43f2-a1ae-4ae80e2b42f7-kube-api-access-zj4ft\") on node \"crc\" DevicePath \"\"" Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.116577 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97a6295d-e613-43f2-a1ae-4ae80e2b42f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97a6295d-e613-43f2-a1ae-4ae80e2b42f7" (UID: "97a6295d-e613-43f2-a1ae-4ae80e2b42f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.216131 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97a6295d-e613-43f2-a1ae-4ae80e2b42f7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.282133 4934 generic.go:334] "Generic (PLEG): container finished" podID="97a6295d-e613-43f2-a1ae-4ae80e2b42f7" containerID="2fe419065b547701ccec2482471f6a79002329ff9b93f456170003aa9f1fe01a" exitCode=0 Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.282219 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-57s9g" Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.282244 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-57s9g" event={"ID":"97a6295d-e613-43f2-a1ae-4ae80e2b42f7","Type":"ContainerDied","Data":"2fe419065b547701ccec2482471f6a79002329ff9b93f456170003aa9f1fe01a"} Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.282290 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-57s9g" event={"ID":"97a6295d-e613-43f2-a1ae-4ae80e2b42f7","Type":"ContainerDied","Data":"f1a6151cd4fc129fb96b02a5c33a9afc7aad09e39b75df088b439470c43a4be4"} Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.282314 4934 scope.go:117] "RemoveContainer" containerID="2fe419065b547701ccec2482471f6a79002329ff9b93f456170003aa9f1fe01a" Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.286832 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="8224ba78a25faf72ee4cb82888cba291e56d3f159515c14eec31cad4a6c0f282" exitCode=0 Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.286873 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"8224ba78a25faf72ee4cb82888cba291e56d3f159515c14eec31cad4a6c0f282"} Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.286899 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da"} Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.326453 4934 scope.go:117] "RemoveContainer" containerID="cae8a9da836a94620299773fbfef723623ca22bb429332c322c4d895bfb43533" Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.359955 4934 scope.go:117] "RemoveContainer" containerID="72dac9d4b98af4f16f1e27fbb67833531667e288fccf7785157ccab6c2e8b987" Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.376621 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-57s9g"] Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.385714 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-57s9g"] Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.391759 4934 scope.go:117] "RemoveContainer" containerID="2fe419065b547701ccec2482471f6a79002329ff9b93f456170003aa9f1fe01a" Dec 27 09:03:16 crc kubenswrapper[4934]: E1227 09:03:16.392221 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fe419065b547701ccec2482471f6a79002329ff9b93f456170003aa9f1fe01a\": container with ID starting with 2fe419065b547701ccec2482471f6a79002329ff9b93f456170003aa9f1fe01a not found: ID does not exist" containerID="2fe419065b547701ccec2482471f6a79002329ff9b93f456170003aa9f1fe01a" Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.392266 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fe419065b547701ccec2482471f6a79002329ff9b93f456170003aa9f1fe01a"} err="failed to get container status \"2fe419065b547701ccec2482471f6a79002329ff9b93f456170003aa9f1fe01a\": rpc error: code = NotFound desc = could not find container \"2fe419065b547701ccec2482471f6a79002329ff9b93f456170003aa9f1fe01a\": container with ID starting with 2fe419065b547701ccec2482471f6a79002329ff9b93f456170003aa9f1fe01a not found: ID does not exist" Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.392295 4934 scope.go:117] "RemoveContainer" containerID="cae8a9da836a94620299773fbfef723623ca22bb429332c322c4d895bfb43533" Dec 27 09:03:16 crc kubenswrapper[4934]: E1227 09:03:16.392878 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cae8a9da836a94620299773fbfef723623ca22bb429332c322c4d895bfb43533\": container with ID starting with cae8a9da836a94620299773fbfef723623ca22bb429332c322c4d895bfb43533 not found: ID does not exist" containerID="cae8a9da836a94620299773fbfef723623ca22bb429332c322c4d895bfb43533" Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.392902 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cae8a9da836a94620299773fbfef723623ca22bb429332c322c4d895bfb43533"} err="failed to get container status \"cae8a9da836a94620299773fbfef723623ca22bb429332c322c4d895bfb43533\": rpc error: code = NotFound desc = could not find container \"cae8a9da836a94620299773fbfef723623ca22bb429332c322c4d895bfb43533\": container with ID starting with cae8a9da836a94620299773fbfef723623ca22bb429332c322c4d895bfb43533 not found: ID does not exist" Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.392915 4934 scope.go:117] "RemoveContainer" containerID="72dac9d4b98af4f16f1e27fbb67833531667e288fccf7785157ccab6c2e8b987" Dec 27 09:03:16 crc kubenswrapper[4934]: E1227 09:03:16.393229 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72dac9d4b98af4f16f1e27fbb67833531667e288fccf7785157ccab6c2e8b987\": container with ID starting with 72dac9d4b98af4f16f1e27fbb67833531667e288fccf7785157ccab6c2e8b987 not found: ID does not exist" containerID="72dac9d4b98af4f16f1e27fbb67833531667e288fccf7785157ccab6c2e8b987" Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.393264 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72dac9d4b98af4f16f1e27fbb67833531667e288fccf7785157ccab6c2e8b987"} err="failed to get container status \"72dac9d4b98af4f16f1e27fbb67833531667e288fccf7785157ccab6c2e8b987\": rpc error: code = NotFound desc = could not find container \"72dac9d4b98af4f16f1e27fbb67833531667e288fccf7785157ccab6c2e8b987\": container with ID starting with 72dac9d4b98af4f16f1e27fbb67833531667e288fccf7785157ccab6c2e8b987 not found: ID does not exist" Dec 27 09:03:16 crc kubenswrapper[4934]: I1227 09:03:16.393287 4934 scope.go:117] "RemoveContainer" containerID="e70b40688ef2f19ae45e5b2903c96bc97f08ed16830328146c09c6c97e7c5c84" Dec 27 09:03:17 crc kubenswrapper[4934]: I1227 09:03:17.484848 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97a6295d-e613-43f2-a1ae-4ae80e2b42f7" path="/var/lib/kubelet/pods/97a6295d-e613-43f2-a1ae-4ae80e2b42f7/volumes" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.358408 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 27 09:03:54 crc kubenswrapper[4934]: E1227 09:03:54.359563 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="175a7c54-e0a9-4f22-8d6d-f74422c2505c" containerName="registry-server" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.359581 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="175a7c54-e0a9-4f22-8d6d-f74422c2505c" containerName="registry-server" Dec 27 09:03:54 crc kubenswrapper[4934]: E1227 09:03:54.359605 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="175a7c54-e0a9-4f22-8d6d-f74422c2505c" containerName="extract-content" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.359613 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="175a7c54-e0a9-4f22-8d6d-f74422c2505c" containerName="extract-content" Dec 27 09:03:54 crc kubenswrapper[4934]: E1227 09:03:54.359634 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97a6295d-e613-43f2-a1ae-4ae80e2b42f7" containerName="registry-server" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.359642 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="97a6295d-e613-43f2-a1ae-4ae80e2b42f7" containerName="registry-server" Dec 27 09:03:54 crc kubenswrapper[4934]: E1227 09:03:54.359655 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97a6295d-e613-43f2-a1ae-4ae80e2b42f7" containerName="extract-utilities" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.359662 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="97a6295d-e613-43f2-a1ae-4ae80e2b42f7" containerName="extract-utilities" Dec 27 09:03:54 crc kubenswrapper[4934]: E1227 09:03:54.359710 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="175a7c54-e0a9-4f22-8d6d-f74422c2505c" containerName="extract-utilities" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.359718 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="175a7c54-e0a9-4f22-8d6d-f74422c2505c" containerName="extract-utilities" Dec 27 09:03:54 crc kubenswrapper[4934]: E1227 09:03:54.359743 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97a6295d-e613-43f2-a1ae-4ae80e2b42f7" containerName="extract-content" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.359749 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="97a6295d-e613-43f2-a1ae-4ae80e2b42f7" containerName="extract-content" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.360058 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="97a6295d-e613-43f2-a1ae-4ae80e2b42f7" containerName="registry-server" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.360074 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="175a7c54-e0a9-4f22-8d6d-f74422c2505c" containerName="registry-server" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.361191 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.363588 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.365198 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vjwww" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.365362 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.365586 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.373917 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.468695 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d38686ee-0d94-4694-845a-e1ee5971af6f-config-data\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.468750 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.468783 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d38686ee-0d94-4694-845a-e1ee5971af6f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.468813 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d38686ee-0d94-4694-845a-e1ee5971af6f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.468848 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d38686ee-0d94-4694-845a-e1ee5971af6f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.469090 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d38686ee-0d94-4694-845a-e1ee5971af6f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.469147 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2pr9\" (UniqueName: \"kubernetes.io/projected/d38686ee-0d94-4694-845a-e1ee5971af6f-kube-api-access-h2pr9\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.469231 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d38686ee-0d94-4694-845a-e1ee5971af6f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.469349 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d38686ee-0d94-4694-845a-e1ee5971af6f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.572000 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d38686ee-0d94-4694-845a-e1ee5971af6f-config-data\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.572150 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.572195 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d38686ee-0d94-4694-845a-e1ee5971af6f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.572255 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d38686ee-0d94-4694-845a-e1ee5971af6f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.572306 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d38686ee-0d94-4694-845a-e1ee5971af6f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.572438 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d38686ee-0d94-4694-845a-e1ee5971af6f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.572465 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2pr9\" (UniqueName: \"kubernetes.io/projected/d38686ee-0d94-4694-845a-e1ee5971af6f-kube-api-access-h2pr9\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.572527 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d38686ee-0d94-4694-845a-e1ee5971af6f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.572555 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d38686ee-0d94-4694-845a-e1ee5971af6f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.573330 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d38686ee-0d94-4694-845a-e1ee5971af6f-config-data\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.574072 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d38686ee-0d94-4694-845a-e1ee5971af6f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.574385 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d38686ee-0d94-4694-845a-e1ee5971af6f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.575535 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d38686ee-0d94-4694-845a-e1ee5971af6f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.575902 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.581107 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d38686ee-0d94-4694-845a-e1ee5971af6f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.582200 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d38686ee-0d94-4694-845a-e1ee5971af6f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.583348 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d38686ee-0d94-4694-845a-e1ee5971af6f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.602918 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2pr9\" (UniqueName: \"kubernetes.io/projected/d38686ee-0d94-4694-845a-e1ee5971af6f-kube-api-access-h2pr9\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.637313 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"tempest-tests-tempest\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " pod="openstack/tempest-tests-tempest" Dec 27 09:03:54 crc kubenswrapper[4934]: I1227 09:03:54.682522 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 27 09:03:55 crc kubenswrapper[4934]: I1227 09:03:55.184801 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 27 09:03:55 crc kubenswrapper[4934]: I1227 09:03:55.780930 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d38686ee-0d94-4694-845a-e1ee5971af6f","Type":"ContainerStarted","Data":"3a51ec0f1172ba43a52c38f4520a3082598a868ea2e3498e9e313f13152a967a"} Dec 27 09:04:27 crc kubenswrapper[4934]: E1227 09:04:27.103952 4934 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 27 09:04:27 crc kubenswrapper[4934]: E1227 09:04:27.106649 4934 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h2pr9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(d38686ee-0d94-4694-845a-e1ee5971af6f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 27 09:04:27 crc kubenswrapper[4934]: E1227 09:04:27.107912 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="d38686ee-0d94-4694-845a-e1ee5971af6f" Dec 27 09:04:27 crc kubenswrapper[4934]: E1227 09:04:27.199715 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="d38686ee-0d94-4694-845a-e1ee5971af6f" Dec 27 09:04:42 crc kubenswrapper[4934]: I1227 09:04:42.028037 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 27 09:04:43 crc kubenswrapper[4934]: I1227 09:04:43.425639 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d38686ee-0d94-4694-845a-e1ee5971af6f","Type":"ContainerStarted","Data":"087703ddc9f57a6c49b83599dfc1da1aa1da6ca6d5fd60592955500f5be1ab77"} Dec 27 09:04:43 crc kubenswrapper[4934]: I1227 09:04:43.442850 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.616174185 podStartE2EDuration="50.442817636s" podCreationTimestamp="2025-12-27 09:03:53 +0000 UTC" firstStartedPulling="2025-12-27 09:03:55.19775216 +0000 UTC m=+4896.018192764" lastFinishedPulling="2025-12-27 09:04:42.024395591 +0000 UTC m=+4942.844836215" observedRunningTime="2025-12-27 09:04:43.442411146 +0000 UTC m=+4944.262851760" watchObservedRunningTime="2025-12-27 09:04:43.442817636 +0000 UTC m=+4944.263258240" Dec 27 09:05:15 crc kubenswrapper[4934]: I1227 09:05:15.329849 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 09:05:15 crc kubenswrapper[4934]: I1227 09:05:15.330330 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 09:05:45 crc kubenswrapper[4934]: I1227 09:05:45.329844 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 09:05:45 crc kubenswrapper[4934]: I1227 09:05:45.330570 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 09:06:15 crc kubenswrapper[4934]: I1227 09:06:15.330393 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 09:06:15 crc kubenswrapper[4934]: I1227 09:06:15.331119 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 09:06:15 crc kubenswrapper[4934]: I1227 09:06:15.331189 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 09:06:15 crc kubenswrapper[4934]: I1227 09:06:15.332463 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 09:06:15 crc kubenswrapper[4934]: I1227 09:06:15.333385 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" gracePeriod=600 Dec 27 09:06:15 crc kubenswrapper[4934]: E1227 09:06:15.481633 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:06:15 crc kubenswrapper[4934]: I1227 09:06:15.587202 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" exitCode=0 Dec 27 09:06:15 crc kubenswrapper[4934]: I1227 09:06:15.587283 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da"} Dec 27 09:06:15 crc kubenswrapper[4934]: I1227 09:06:15.588048 4934 scope.go:117] "RemoveContainer" containerID="8224ba78a25faf72ee4cb82888cba291e56d3f159515c14eec31cad4a6c0f282" Dec 27 09:06:15 crc kubenswrapper[4934]: I1227 09:06:15.588675 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:06:15 crc kubenswrapper[4934]: E1227 09:06:15.589502 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:06:26 crc kubenswrapper[4934]: I1227 09:06:26.467924 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:06:26 crc kubenswrapper[4934]: E1227 09:06:26.468978 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:06:37 crc kubenswrapper[4934]: I1227 09:06:37.468252 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:06:37 crc kubenswrapper[4934]: E1227 09:06:37.469156 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:06:52 crc kubenswrapper[4934]: I1227 09:06:52.468765 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:06:52 crc kubenswrapper[4934]: E1227 09:06:52.469558 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:07:07 crc kubenswrapper[4934]: I1227 09:07:07.467835 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:07:07 crc kubenswrapper[4934]: E1227 09:07:07.468728 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:07:20 crc kubenswrapper[4934]: I1227 09:07:20.467477 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:07:20 crc kubenswrapper[4934]: E1227 09:07:20.468155 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:07:23 crc kubenswrapper[4934]: I1227 09:07:23.394246 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hrs9g"] Dec 27 09:07:23 crc kubenswrapper[4934]: I1227 09:07:23.399648 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hrs9g" Dec 27 09:07:23 crc kubenswrapper[4934]: I1227 09:07:23.505008 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6rkd\" (UniqueName: \"kubernetes.io/projected/73bc5c15-5c9d-49ae-bcd9-853df82fabbf-kube-api-access-j6rkd\") pod \"community-operators-hrs9g\" (UID: \"73bc5c15-5c9d-49ae-bcd9-853df82fabbf\") " pod="openshift-marketplace/community-operators-hrs9g" Dec 27 09:07:23 crc kubenswrapper[4934]: I1227 09:07:23.505146 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73bc5c15-5c9d-49ae-bcd9-853df82fabbf-utilities\") pod \"community-operators-hrs9g\" (UID: \"73bc5c15-5c9d-49ae-bcd9-853df82fabbf\") " pod="openshift-marketplace/community-operators-hrs9g" Dec 27 09:07:23 crc kubenswrapper[4934]: I1227 09:07:23.505239 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73bc5c15-5c9d-49ae-bcd9-853df82fabbf-catalog-content\") pod \"community-operators-hrs9g\" (UID: \"73bc5c15-5c9d-49ae-bcd9-853df82fabbf\") " pod="openshift-marketplace/community-operators-hrs9g" Dec 27 09:07:23 crc kubenswrapper[4934]: I1227 09:07:23.580804 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hrs9g"] Dec 27 09:07:23 crc kubenswrapper[4934]: I1227 09:07:23.608935 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6rkd\" (UniqueName: \"kubernetes.io/projected/73bc5c15-5c9d-49ae-bcd9-853df82fabbf-kube-api-access-j6rkd\") pod \"community-operators-hrs9g\" (UID: \"73bc5c15-5c9d-49ae-bcd9-853df82fabbf\") " pod="openshift-marketplace/community-operators-hrs9g" Dec 27 09:07:23 crc kubenswrapper[4934]: I1227 09:07:23.609041 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73bc5c15-5c9d-49ae-bcd9-853df82fabbf-utilities\") pod \"community-operators-hrs9g\" (UID: \"73bc5c15-5c9d-49ae-bcd9-853df82fabbf\") " pod="openshift-marketplace/community-operators-hrs9g" Dec 27 09:07:23 crc kubenswrapper[4934]: I1227 09:07:23.609144 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73bc5c15-5c9d-49ae-bcd9-853df82fabbf-catalog-content\") pod \"community-operators-hrs9g\" (UID: \"73bc5c15-5c9d-49ae-bcd9-853df82fabbf\") " pod="openshift-marketplace/community-operators-hrs9g" Dec 27 09:07:23 crc kubenswrapper[4934]: I1227 09:07:23.611404 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73bc5c15-5c9d-49ae-bcd9-853df82fabbf-catalog-content\") pod \"community-operators-hrs9g\" (UID: \"73bc5c15-5c9d-49ae-bcd9-853df82fabbf\") " pod="openshift-marketplace/community-operators-hrs9g" Dec 27 09:07:23 crc kubenswrapper[4934]: I1227 09:07:23.611879 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73bc5c15-5c9d-49ae-bcd9-853df82fabbf-utilities\") pod \"community-operators-hrs9g\" (UID: \"73bc5c15-5c9d-49ae-bcd9-853df82fabbf\") " pod="openshift-marketplace/community-operators-hrs9g" Dec 27 09:07:23 crc kubenswrapper[4934]: I1227 09:07:23.637792 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6rkd\" (UniqueName: \"kubernetes.io/projected/73bc5c15-5c9d-49ae-bcd9-853df82fabbf-kube-api-access-j6rkd\") pod \"community-operators-hrs9g\" (UID: \"73bc5c15-5c9d-49ae-bcd9-853df82fabbf\") " pod="openshift-marketplace/community-operators-hrs9g" Dec 27 09:07:23 crc kubenswrapper[4934]: I1227 09:07:23.728854 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hrs9g" Dec 27 09:07:24 crc kubenswrapper[4934]: I1227 09:07:24.613953 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hrs9g"] Dec 27 09:07:25 crc kubenswrapper[4934]: I1227 09:07:25.445865 4934 generic.go:334] "Generic (PLEG): container finished" podID="73bc5c15-5c9d-49ae-bcd9-853df82fabbf" containerID="e0f3661839b28d617baf5eea31c844302cea31ec2a1859bff75b55bd5dfa0bd3" exitCode=0 Dec 27 09:07:25 crc kubenswrapper[4934]: I1227 09:07:25.445866 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrs9g" event={"ID":"73bc5c15-5c9d-49ae-bcd9-853df82fabbf","Type":"ContainerDied","Data":"e0f3661839b28d617baf5eea31c844302cea31ec2a1859bff75b55bd5dfa0bd3"} Dec 27 09:07:25 crc kubenswrapper[4934]: I1227 09:07:25.446247 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrs9g" event={"ID":"73bc5c15-5c9d-49ae-bcd9-853df82fabbf","Type":"ContainerStarted","Data":"e3a9ef9e90cf6f45c7368ba388808bfbff8b7672e6a124fbda7635c6f8dc1365"} Dec 27 09:07:26 crc kubenswrapper[4934]: I1227 09:07:26.456538 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrs9g" event={"ID":"73bc5c15-5c9d-49ae-bcd9-853df82fabbf","Type":"ContainerStarted","Data":"ed04061c0c586c29911ef265fb4612200b6233b0f439ff8b3b2c07885d724f48"} Dec 27 09:07:28 crc kubenswrapper[4934]: I1227 09:07:28.527012 4934 generic.go:334] "Generic (PLEG): container finished" podID="73bc5c15-5c9d-49ae-bcd9-853df82fabbf" containerID="ed04061c0c586c29911ef265fb4612200b6233b0f439ff8b3b2c07885d724f48" exitCode=0 Dec 27 09:07:28 crc kubenswrapper[4934]: I1227 09:07:28.527173 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrs9g" event={"ID":"73bc5c15-5c9d-49ae-bcd9-853df82fabbf","Type":"ContainerDied","Data":"ed04061c0c586c29911ef265fb4612200b6233b0f439ff8b3b2c07885d724f48"} Dec 27 09:07:30 crc kubenswrapper[4934]: I1227 09:07:30.551707 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrs9g" event={"ID":"73bc5c15-5c9d-49ae-bcd9-853df82fabbf","Type":"ContainerStarted","Data":"23bb638c28ceb00176e086fce835249595930f3094fbf84d7920bc0e0e3d1de1"} Dec 27 09:07:30 crc kubenswrapper[4934]: I1227 09:07:30.580833 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hrs9g" podStartSLOduration=4.025730455 podStartE2EDuration="7.578466676s" podCreationTimestamp="2025-12-27 09:07:23 +0000 UTC" firstStartedPulling="2025-12-27 09:07:25.447816495 +0000 UTC m=+5106.268257099" lastFinishedPulling="2025-12-27 09:07:29.000552716 +0000 UTC m=+5109.820993320" observedRunningTime="2025-12-27 09:07:30.571075012 +0000 UTC m=+5111.391515626" watchObservedRunningTime="2025-12-27 09:07:30.578466676 +0000 UTC m=+5111.398907260" Dec 27 09:07:31 crc kubenswrapper[4934]: I1227 09:07:31.468198 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:07:31 crc kubenswrapper[4934]: E1227 09:07:31.468527 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:07:33 crc kubenswrapper[4934]: I1227 09:07:33.729950 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hrs9g" Dec 27 09:07:33 crc kubenswrapper[4934]: I1227 09:07:33.730691 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hrs9g" Dec 27 09:07:34 crc kubenswrapper[4934]: I1227 09:07:34.798050 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-hrs9g" podUID="73bc5c15-5c9d-49ae-bcd9-853df82fabbf" containerName="registry-server" probeResult="failure" output=< Dec 27 09:07:34 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 09:07:34 crc kubenswrapper[4934]: > Dec 27 09:07:41 crc kubenswrapper[4934]: I1227 09:07:41.589549 4934 patch_prober.go:28] interesting pod/controller-manager-7967777f79-kzj4d container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.83:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:41 crc kubenswrapper[4934]: I1227 09:07:41.595076 4934 patch_prober.go:28] interesting pod/controller-manager-7967777f79-kzj4d container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.83:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:41 crc kubenswrapper[4934]: I1227 09:07:41.595535 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" podUID="2a688ece-0e00-4241-b7dd-9dde25f0eaf1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.83:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:41 crc kubenswrapper[4934]: I1227 09:07:41.595176 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" podUID="2a688ece-0e00-4241-b7dd-9dde25f0eaf1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.83:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:41 crc kubenswrapper[4934]: I1227 09:07:41.806214 4934 patch_prober.go:28] interesting pod/logging-loki-distributor-5f678c8dd6-nqqn7 container/loki-distributor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.50:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:41 crc kubenswrapper[4934]: I1227 09:07:41.806299 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" podUID="5864b138-b8ef-405e-9ed9-be39f8e13e8d" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.50:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:42 crc kubenswrapper[4934]: I1227 09:07:42.503012 4934 patch_prober.go:28] interesting pod/logging-loki-querier-76788598db-6wrwh container/loki-querier namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.51:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:42 crc kubenswrapper[4934]: I1227 09:07:42.503662 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" podUID="ae25e4b9-7a2a-41ff-9944-d6c45603cbec" containerName="loki-querier" probeResult="failure" output="Get \"https://10.217.0.51:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:42 crc kubenswrapper[4934]: I1227 09:07:42.505404 4934 patch_prober.go:28] interesting pod/thanos-querier-8467887986-svb7h container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.71:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:42 crc kubenswrapper[4934]: I1227 09:07:42.505485 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-8467887986-svb7h" podUID="a07fdfc4-584a-4a39-b87f-ee78f9cac687" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.71:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:42 crc kubenswrapper[4934]: I1227 09:07:42.512889 4934 patch_prober.go:28] interesting pod/logging-loki-query-frontend-69d9546745-4b9rt container/loki-query-frontend namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.52:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:42 crc kubenswrapper[4934]: I1227 09:07:42.512966 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" podUID="b9e3cc00-7197-4c75-a732-136e65d893c8" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.52:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:42 crc kubenswrapper[4934]: I1227 09:07:42.640826 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="ca95a4ef-282d-4c26-983f-d7d00f90443c" containerName="galera" probeResult="failure" output="command timed out" Dec 27 09:07:42 crc kubenswrapper[4934]: I1227 09:07:42.640832 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="ca95a4ef-282d-4c26-983f-d7d00f90443c" containerName="galera" probeResult="failure" output="command timed out" Dec 27 09:07:42 crc kubenswrapper[4934]: I1227 09:07:42.805328 4934 patch_prober.go:28] interesting pod/logging-loki-distributor-5f678c8dd6-nqqn7 container/loki-distributor namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.50:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:42 crc kubenswrapper[4934]: I1227 09:07:42.805415 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" podUID="5864b138-b8ef-405e-9ed9-be39f8e13e8d" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.50:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:42 crc kubenswrapper[4934]: I1227 09:07:42.836771 4934 patch_prober.go:28] interesting pod/nmstate-webhook-f8fb84555-fwjdt container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.86:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:42 crc kubenswrapper[4934]: I1227 09:07:42.836848 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt" podUID="415e9aa6-5a7a-4d93-924d-0213c9a5ca4d" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.86:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:42 crc kubenswrapper[4934]: I1227 09:07:42.984200 4934 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:42 crc kubenswrapper[4934]: I1227 09:07:42.984280 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="642f3213-8063-4ade-800e-4e05ba16d642" containerName="loki-ingester" probeResult="failure" output="Get \"https://10.217.0.55:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:42 crc kubenswrapper[4934]: I1227 09:07:42.991510 4934 patch_prober.go:28] interesting pod/route-controller-manager-6bd9bdbfbf-x5vvz container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.77:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:42 crc kubenswrapper[4934]: I1227 09:07:42.991560 4934 patch_prober.go:28] interesting pod/route-controller-manager-6bd9bdbfbf-x5vvz container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.77:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:42 crc kubenswrapper[4934]: I1227 09:07:42.991573 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" podUID="45f4ca46-4041-4960-8c05-a20bd2c90b90" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.77:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:42 crc kubenswrapper[4934]: I1227 09:07:42.991615 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" podUID="45f4ca46-4041-4960-8c05-a20bd2c90b90" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.77:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:43 crc kubenswrapper[4934]: I1227 09:07:43.065491 4934 patch_prober.go:28] interesting pod/logging-loki-compactor-0 container/loki-compactor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.56:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:43 crc kubenswrapper[4934]: I1227 09:07:43.065556 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-compactor-0" podUID="e6de9198-9266-43e8-b760-e21f0afa885e" containerName="loki-compactor" probeResult="failure" output="Get \"https://10.217.0.56:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:43 crc kubenswrapper[4934]: I1227 09:07:43.133907 4934 patch_prober.go:28] interesting pod/logging-loki-index-gateway-0 container/loki-index-gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:43 crc kubenswrapper[4934]: I1227 09:07:43.133973 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-index-gateway-0" podUID="01f91788-9f82-4508-906b-1e98c9e05c2c" containerName="loki-index-gateway" probeResult="failure" output="Get \"https://10.217.0.57:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:43 crc kubenswrapper[4934]: I1227 09:07:43.503564 4934 patch_prober.go:28] interesting pod/logging-loki-querier-76788598db-6wrwh container/loki-querier namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.51:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:43 crc kubenswrapper[4934]: I1227 09:07:43.503855 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" podUID="ae25e4b9-7a2a-41ff-9944-d6c45603cbec" containerName="loki-querier" probeResult="failure" output="Get \"https://10.217.0.51:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:43 crc kubenswrapper[4934]: I1227 09:07:43.512725 4934 patch_prober.go:28] interesting pod/logging-loki-query-frontend-69d9546745-4b9rt container/loki-query-frontend namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.52:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:43 crc kubenswrapper[4934]: I1227 09:07:43.512766 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" podUID="b9e3cc00-7197-4c75-a732-136e65d893c8" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.52:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:43 crc kubenswrapper[4934]: I1227 09:07:43.548245 4934 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5jzhr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.63:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:43 crc kubenswrapper[4934]: I1227 09:07:43.548245 4934 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5jzhr container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.63:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:43 crc kubenswrapper[4934]: I1227 09:07:43.548298 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" podUID="a4612dd4-0e0d-4c38-8da5-8bc30dee7c12" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.63:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:43 crc kubenswrapper[4934]: I1227 09:07:43.548309 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" podUID="a4612dd4-0e0d-4c38-8da5-8bc30dee7c12" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.63:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:43 crc kubenswrapper[4934]: I1227 09:07:43.682632 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4" podUID="935a4c26-2678-4ead-87ee-1297850d89ff" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.102:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:43 crc kubenswrapper[4934]: I1227 09:07:43.723300 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" podUID="dc4ddef4-e9a5-4934-846e-1d1bcfe9f223" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.103:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:43 crc kubenswrapper[4934]: I1227 09:07:43.723420 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx" podUID="fc1c7052-d894-4fa1-b6bb-951cad9e32e0" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.104:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:43 crc kubenswrapper[4934]: I1227 09:07:43.765292 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/glance-operator-controller-manager-64fb555449-tbmpv" podUID="3df60bce-0768-4949-8b8d-3bb28566cd4a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:43 crc kubenswrapper[4934]: I1227 09:07:43.806489 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" podUID="77029090-da44-403b-a20f-d72105281956" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:43 crc kubenswrapper[4934]: I1227 09:07:43.856422 4934 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Liveness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:43 crc kubenswrapper[4934]: I1227 09:07:43.856560 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:43 crc kubenswrapper[4934]: I1227 09:07:43.982139 4934 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.55:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:43 crc kubenswrapper[4934]: I1227 09:07:43.982206 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-ingester-0" podUID="642f3213-8063-4ade-800e-4e05ba16d642" containerName="loki-ingester" probeResult="failure" output="Get \"https://10.217.0.55:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.065434 4934 patch_prober.go:28] interesting pod/logging-loki-compactor-0 container/loki-compactor namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.56:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.065522 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-compactor-0" podUID="e6de9198-9266-43e8-b760-e21f0afa885e" containerName="loki-compactor" probeResult="failure" output="Get \"https://10.217.0.56:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.133068 4934 patch_prober.go:28] interesting pod/logging-loki-index-gateway-0 container/loki-index-gateway namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.57:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.133151 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-index-gateway-0" podUID="01f91788-9f82-4508-906b-1e98c9e05c2c" containerName="loki-index-gateway" probeResult="failure" output="Get \"https://10.217.0.57:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.221453 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" podUID="718122b6-c9f4-4d6f-92b5-6862c15f4ee6" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.110:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.221496 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/manila-operator-controller-manager-7cc599445b-dlwjl" podUID="37d56061-eab6-43e5-8dec-f59220da5d0d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.221545 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59" podUID="828ebb68-326b-4c32-bc0a-d7f258b45ebd" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.312375 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" podUID="1c0960f0-ef55-479a-b375-a09e69254743" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.312495 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" podUID="d090706d-34a7-4c72-9cb9-ad0601db35a6" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.395336 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" podUID="e2db190a-b455-44fc-a43e-2677e5af27b2" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.108:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.395369 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" podUID="59d65d09-c15c-49d9-8ee7-472ffe0dda55" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.436281 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg" podUID="4f825c5b-90ef-4b56-a759-d7ccbb312bed" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.515851 4934 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-2n6w2 container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.515895 4934 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-2n6w2 container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.515918 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" podUID="964b7818-b04e-4c6d-8ad0-5a108513be73" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.515960 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" podUID="964b7818-b04e-4c6d-8ad0-5a108513be73" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.567287 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj" podUID="f2d5e345-1bd9-4313-bdae-7637cb7ed944" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.577742 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" podUID="f1cce183-e0a8-4cf1-854e-7a60986f194f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.633395 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" podUID="b5ea792c-b43f-4221-b7bb-aba3421de3d8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.639765 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="727f882e-fa63-46e9-9cd5-9781e472f2ef" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.640955 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-northd-0" podUID="727f882e-fa63-46e9-9cd5-9781e472f2ef" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.642037 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="40c7cd23-14a7-4b30-9639-84d83ba74fe3" containerName="prometheus" probeResult="failure" output="command timed out" Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.740308 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" podUID="35d6eb8d-3797-432c-a5ea-65f31adf7262" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:44 crc kubenswrapper[4934]: I1227 09:07:44.830324 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf" podUID="e44e23aa-f466-41cb-b855-d0bdc25cf05b" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:45 crc kubenswrapper[4934]: I1227 09:07:45.145277 4934 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:45 crc kubenswrapper[4934]: I1227 09:07:45.145395 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:45 crc kubenswrapper[4934]: I1227 09:07:45.239754 4934 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5fhq7 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:45 crc kubenswrapper[4934]: I1227 09:07:45.239791 4934 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5fhq7 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:45 crc kubenswrapper[4934]: I1227 09:07:45.239823 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" podUID="4c71cabc-c13f-4ae8-a704-10efee323e35" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:45 crc kubenswrapper[4934]: I1227 09:07:45.239823 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" podUID="4c71cabc-c13f-4ae8-a704-10efee323e35" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:45 crc kubenswrapper[4934]: I1227 09:07:45.331328 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-687f57d79b-5kwg4" podUID="dab3e53b-cbb2-4423-a8e3-432972ecedc3" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.44:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:45 crc kubenswrapper[4934]: I1227 09:07:45.488125 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:07:45 crc kubenswrapper[4934]: E1227 09:07:45.490840 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:07:45 crc kubenswrapper[4934]: I1227 09:07:45.643756 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="40c7cd23-14a7-4b30-9639-84d83ba74fe3" containerName="prometheus" probeResult="failure" output="command timed out" Dec 27 09:07:45 crc kubenswrapper[4934]: I1227 09:07:45.753371 4934 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-lnknm container/operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:45 crc kubenswrapper[4934]: I1227 09:07:45.753517 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" podUID="4575d285-879f-4cc3-a9b9-406dc96e852d" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:45 crc kubenswrapper[4934]: I1227 09:07:45.753402 4934 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-lnknm container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:45 crc kubenswrapper[4934]: I1227 09:07:45.753656 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" podUID="4575d285-879f-4cc3-a9b9-406dc96e852d" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:45 crc kubenswrapper[4934]: I1227 09:07:45.886255 4934 patch_prober.go:28] interesting pod/perses-operator-5bf474d74f-tx6z5 container/perses-operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.13:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:45 crc kubenswrapper[4934]: I1227 09:07:45.886382 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" podUID="f97e8d86-1e9e-4c5a-a164-2600193c576b" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.13:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:45 crc kubenswrapper[4934]: I1227 09:07:45.927470 4934 patch_prober.go:28] interesting pod/perses-operator-5bf474d74f-tx6z5 container/perses-operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.13:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:45 crc kubenswrapper[4934]: I1227 09:07:45.927555 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" podUID="f97e8d86-1e9e-4c5a-a164-2600193c576b" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.13:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:45 crc kubenswrapper[4934]: I1227 09:07:45.984989 4934 patch_prober.go:28] interesting pod/console-74cc7c486-2dpkp container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.138:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:45 crc kubenswrapper[4934]: I1227 09:07:45.985118 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-74cc7c486-2dpkp" podUID="e0978ad2-1582-49d4-b239-ff929dafb117" containerName="console" probeResult="failure" output="Get \"https://10.217.0.138:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:46 crc kubenswrapper[4934]: I1227 09:07:46.096359 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc" podUID="b461380d-6573-4898-9bc8-c82ceba822d8" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.101:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:46 crc kubenswrapper[4934]: I1227 09:07:46.096391 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc" podUID="b461380d-6573-4898-9bc8-c82ceba822d8" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.101:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:46 crc kubenswrapper[4934]: I1227 09:07:46.511840 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="4e39152f-3592-4907-9b06-d13074aeb3c4" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.5:8081/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:46 crc kubenswrapper[4934]: I1227 09:07:46.511892 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="4e39152f-3592-4907-9b06-d13074aeb3c4" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.5:8080/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.064236 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" podUID="fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.93:8080/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.428379 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" podUID="df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.94:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.428360 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" podUID="df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.94:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.452588 4934 patch_prober.go:28] interesting pod/thanos-querier-8467887986-svb7h container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.71:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.452668 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-8467887986-svb7h" podUID="a07fdfc4-584a-4a39-b87f-ee78f9cac687" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.71:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.559237 4934 patch_prober.go:28] interesting pod/logging-loki-gateway-74dbf44c45-qd69g container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.559337 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" podUID="47cbee2e-c325-4a8d-a26d-94bced60ee78" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.559263 4934 patch_prober.go:28] interesting pod/logging-loki-gateway-74dbf44c45-qd69g container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.559511 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" podUID="47cbee2e-c325-4a8d-a26d-94bced60ee78" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.53:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.640887 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.671389 4934 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-dwv8c container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.671788 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" podUID="901df0a0-1dae-41aa-8d29-0981524d9312" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.743612 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="e26e3510-3438-4cb9-8d00-f2d0fdf33ef5" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.171:9090/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.743702 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/prometheus-metric-storage-0" podUID="e26e3510-3438-4cb9-8d00-f2d0fdf33ef5" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.171:9090/-/healthy\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.833277 4934 patch_prober.go:28] interesting pod/downloads-7954f5f757-8msj4 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.833339 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-8msj4" podUID="a0d34185-bf08-456c-a5fc-3f0852d35a91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.876726 4934 patch_prober.go:28] interesting pod/downloads-7954f5f757-8msj4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.876797 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8msj4" podUID="a0d34185-bf08-456c-a5fc-3f0852d35a91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.933096 4934 patch_prober.go:28] interesting pod/logging-loki-gateway-74dbf44c45-4ll7v container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.933174 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" podUID="eac5b26c-d598-4c68-95b5-e2583b456af9" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.933199 4934 patch_prober.go:28] interesting pod/logging-loki-gateway-74dbf44c45-4ll7v container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:47 crc kubenswrapper[4934]: I1227 09:07:47.933261 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" podUID="eac5b26c-d598-4c68-95b5-e2583b456af9" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.54:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.007424 4934 patch_prober.go:28] interesting pod/console-operator-58897d9998-jckgt container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.007450 4934 patch_prober.go:28] interesting pod/console-operator-58897d9998-jckgt container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.007491 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-jckgt" podUID="753c599b-a401-481e-a5b3-e2dd6ed2c4e7" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.007513 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-jckgt" podUID="753c599b-a401-481e-a5b3-e2dd6ed2c4e7" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.176411 4934 patch_prober.go:28] interesting pod/image-registry-66df7c8f76-vkfr4 container/registry namespace/openshift-image-registry: Readiness probe status=failure output="Get \"https://10.217.0.78:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.176494 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" podUID="21dab295-f1e7-4fee-a281-251122757a57" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.78:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.179256 4934 patch_prober.go:28] interesting pod/image-registry-66df7c8f76-vkfr4 container/registry namespace/openshift-image-registry: Liveness probe status=failure output="Get \"https://10.217.0.78:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.179311 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" podUID="21dab295-f1e7-4fee-a281-251122757a57" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.78:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.239293 4934 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5fhq7 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.239352 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" podUID="4c71cabc-c13f-4ae8-a704-10efee323e35" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.239523 4934 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5fhq7 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.239581 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" podUID="4c71cabc-c13f-4ae8-a704-10efee323e35" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.375401 4934 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-5cc9p container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.375617 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5bsq9" podUID="0134c36f-c43b-4830-848a-3a8690957ee1" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.95:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.375766 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5bsq9" podUID="0134c36f-c43b-4830-848a-3a8690957ee1" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.95:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.385429 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" podUID="af023fee-eae1-427a-8d70-487588422316" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.375562 4934 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-5cc9p container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.385523 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" podUID="af023fee-eae1-427a-8d70-487588422316" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.436312 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/controller-5bddd4b946-7c5ls" podUID="a8825499-558f-4b31-b09b-0dcc92662f65" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.96:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.477346 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/controller-5bddd4b946-7c5ls" podUID="a8825499-558f-4b31-b09b-0dcc92662f65" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.96:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.643793 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-hrs9g" podUID="73bc5c15-5c9d-49ae-bcd9-853df82fabbf" containerName="registry-server" probeResult="failure" output="command timed out" Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.674213 4934 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-r6jld container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.674301 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" podUID="4ea539b6-ffe0-4393-af8d-62579266697c" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.674314 4934 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-r6jld container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.674382 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" podUID="4ea539b6-ffe0-4393-af8d-62579266697c" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.720273 4934 patch_prober.go:28] interesting pod/router-default-5444994796-cpts5 container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.720738 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-cpts5" podUID="ac1de940-6952-466a-be96-52c43b1a90a2" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.762247 4934 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-t6q48 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.762310 4934 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-t6q48 container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.762375 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" podUID="873d8371-2f45-409a-b18d-f4c7dbdabeea" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.762310 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" podUID="873d8371-2f45-409a-b18d-f4c7dbdabeea" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.762428 4934 patch_prober.go:28] interesting pod/router-default-5444994796-cpts5 container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.762445 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-cpts5" podUID="ac1de940-6952-466a-be96-52c43b1a90a2" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.954390 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-v2lll" podUID="c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:48 crc kubenswrapper[4934]: I1227 09:07:48.995410 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-v2lll" podUID="c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:49 crc kubenswrapper[4934]: I1227 09:07:49.077376 4934 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-c2tqv container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:49 crc kubenswrapper[4934]: I1227 09:07:49.077434 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-c2tqv" podUID="85b3562e-4570-4b9b-baa5-a20fe55f13ca" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:49 crc kubenswrapper[4934]: I1227 09:07:49.077505 4934 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-c2tqv container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:49 crc kubenswrapper[4934]: I1227 09:07:49.077522 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-c2tqv" podUID="85b3562e-4570-4b9b-baa5-a20fe55f13ca" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:49 crc kubenswrapper[4934]: I1227 09:07:49.077742 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-v2lll" podUID="c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:49 crc kubenswrapper[4934]: I1227 09:07:49.214770 4934 patch_prober.go:28] interesting pod/oauth-openshift-fd9565bb5-mkr88 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.72:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:49 crc kubenswrapper[4934]: I1227 09:07:49.214820 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" podUID="772a3e80-411a-4c15-9e1c-b702392e13af" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.72:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:49 crc kubenswrapper[4934]: I1227 09:07:49.214837 4934 patch_prober.go:28] interesting pod/oauth-openshift-fd9565bb5-mkr88 container/oauth-openshift namespace/openshift-authentication: Liveness probe status=failure output="Get \"https://10.217.0.72:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:49 crc kubenswrapper[4934]: I1227 09:07:49.214908 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" podUID="772a3e80-411a-4c15-9e1c-b702392e13af" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.72:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:49 crc kubenswrapper[4934]: I1227 09:07:49.641252 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-handler-kbxfg" podUID="224f9a57-4665-4c0f-865c-ac2267f69b7e" containerName="nmstate-handler" probeResult="failure" output="command timed out" Dec 27 09:07:49 crc kubenswrapper[4934]: I1227 09:07:49.641252 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="727f882e-fa63-46e9-9cd5-9781e472f2ef" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 27 09:07:49 crc kubenswrapper[4934]: I1227 09:07:49.641330 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-northd-0" podUID="727f882e-fa63-46e9-9cd5-9781e472f2ef" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 27 09:07:49 crc kubenswrapper[4934]: I1227 09:07:49.956269 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-4jwwj" podUID="ea01c7df-c5aa-4c86-800b-b6aebafeb8fd" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:49 crc kubenswrapper[4934]: I1227 09:07:49.956270 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-4jwwj" podUID="ea01c7df-c5aa-4c86-800b-b6aebafeb8fd" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:50 crc kubenswrapper[4934]: I1227 09:07:50.643173 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="40c7cd23-14a7-4b30-9639-84d83ba74fe3" containerName="prometheus" probeResult="failure" output="command timed out" Dec 27 09:07:50 crc kubenswrapper[4934]: I1227 09:07:50.644637 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="40c7cd23-14a7-4b30-9639-84d83ba74fe3" containerName="prometheus" probeResult="failure" output="command timed out" Dec 27 09:07:50 crc kubenswrapper[4934]: I1227 09:07:50.740959 4934 patch_prober.go:28] interesting pod/metrics-server-76d5d9b996-jzgqz container/metrics-server namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.74:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:50 crc kubenswrapper[4934]: I1227 09:07:50.741352 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" podUID="169cb9e6-0d57-425d-a51d-a3b0e1556db6" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.74:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:50 crc kubenswrapper[4934]: I1227 09:07:50.740965 4934 patch_prober.go:28] interesting pod/metrics-server-76d5d9b996-jzgqz container/metrics-server namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.74:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:50 crc kubenswrapper[4934]: I1227 09:07:50.741432 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" podUID="169cb9e6-0d57-425d-a51d-a3b0e1556db6" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.74:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.247694 4934 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5fhq7 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.247745 4934 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5fhq7 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.247768 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" podUID="4c71cabc-c13f-4ae8-a704-10efee323e35" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.247822 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" podUID="4c71cabc-c13f-4ae8-a704-10efee323e35" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.248594 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.248636 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.250509 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="openshift-config-operator" containerStatusID={"Type":"cri-o","ID":"b631e608d82ae75f36a5a9fa3ea7f2bfcda667d3a9703769b16b2e70e44081be"} pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" containerMessage="Container openshift-config-operator failed liveness probe, will be restarted" Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.251237 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" podUID="4c71cabc-c13f-4ae8-a704-10efee323e35" containerName="openshift-config-operator" containerID="cri-o://b631e608d82ae75f36a5a9fa3ea7f2bfcda667d3a9703769b16b2e70e44081be" gracePeriod=30 Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.264435 4934 patch_prober.go:28] interesting pod/monitoring-plugin-7d4f6975b-p9h8q container/monitoring-plugin namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.75:9443/health\": context deadline exceeded" start-of-body= Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.264481 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/monitoring-plugin-7d4f6975b-p9h8q" podUID="b78888a7-451e-43a3-86ea-b13c035afab1" containerName="monitoring-plugin" probeResult="failure" output="Get \"https://10.217.0.75:9443/health\": context deadline exceeded" Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.445628 4934 patch_prober.go:28] interesting pod/loki-operator-controller-manager-77c7786d5-gz9d9 container/manager namespace/openshift-operators-redhat: Readiness probe status=failure output="Get \"http://10.217.0.47:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.445685 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" podUID="00799e47-9b94-49bf-8bd1-d6bb5036285a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.47:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.480399 4934 patch_prober.go:28] interesting pod/controller-manager-7967777f79-kzj4d container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.83:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.483212 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" podUID="2a688ece-0e00-4241-b7dd-9dde25f0eaf1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.83:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.481567 4934 patch_prober.go:28] interesting pod/controller-manager-7967777f79-kzj4d container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.83:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.483289 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" podUID="2a688ece-0e00-4241-b7dd-9dde25f0eaf1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.83:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.641035 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="d03cf582-2192-4a10-b21e-10befebde419" containerName="galera" probeResult="failure" output="command timed out" Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.641048 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="d03cf582-2192-4a10-b21e-10befebde419" containerName="galera" probeResult="failure" output="command timed out" Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.643316 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-index-v7q5s" podUID="a41afe73-59e9-4cbe-b2e7-1f755767daf5" containerName="registry-server" probeResult="failure" output="command timed out" Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.646252 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-operators-qtq22" podUID="2ee60051-8eec-4da3-ae78-b409c5cdf9f3" containerName="registry-server" probeResult="failure" output="command timed out" Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.646342 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-index-v7q5s" podUID="a41afe73-59e9-4cbe-b2e7-1f755767daf5" containerName="registry-server" probeResult="failure" output="command timed out" Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.649573 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-operators-qtq22" podUID="2ee60051-8eec-4da3-ae78-b409c5cdf9f3" containerName="registry-server" probeResult="failure" output="command timed out" Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.806444 4934 patch_prober.go:28] interesting pod/logging-loki-distributor-5f678c8dd6-nqqn7 container/loki-distributor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.50:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:51 crc kubenswrapper[4934]: I1227 09:07:51.806506 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" podUID="5864b138-b8ef-405e-9ed9-be39f8e13e8d" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.50:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.068120 4934 patch_prober.go:28] interesting pod/logging-loki-query-frontend-69d9546745-4b9rt container/loki-query-frontend namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.52:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.068501 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" podUID="b9e3cc00-7197-4c75-a732-136e65d893c8" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.52:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.254400 4934 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5fhq7 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.254517 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" podUID="4c71cabc-c13f-4ae8-a704-10efee323e35" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.268504 4934 patch_prober.go:28] interesting pod/logging-loki-querier-76788598db-6wrwh container/loki-querier namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.51:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.268587 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" podUID="ae25e4b9-7a2a-41ff-9944-d6c45603cbec" containerName="loki-querier" probeResult="failure" output="Get \"https://10.217.0.51:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.397860 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-marketplace-7rs6w" podUID="ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3" containerName="registry-server" probeResult="failure" output=< Dec 27 09:07:52 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 09:07:52 crc kubenswrapper[4934]: > Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.397895 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-sz4sq" podUID="72569b30-4bc6-4f58-b62e-c1a0ef0961d6" containerName="registry-server" probeResult="failure" output=< Dec 27 09:07:52 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 09:07:52 crc kubenswrapper[4934]: > Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.399189 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/certified-operators-lbfd2" podUID="0661ef0e-af9d-41ba-8059-2e878b11b042" containerName="registry-server" probeResult="failure" output=< Dec 27 09:07:52 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 09:07:52 crc kubenswrapper[4934]: > Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.451909 4934 patch_prober.go:28] interesting pod/thanos-querier-8467887986-svb7h container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.71:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.452014 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-8467887986-svb7h" podUID="a07fdfc4-584a-4a39-b87f-ee78f9cac687" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.71:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.559565 4934 patch_prober.go:28] interesting pod/logging-loki-gateway-74dbf44c45-qd69g container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.559590 4934 patch_prober.go:28] interesting pod/logging-loki-gateway-74dbf44c45-qd69g container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.560031 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" podUID="47cbee2e-c325-4a8d-a26d-94bced60ee78" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.53:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.559952 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" podUID="47cbee2e-c325-4a8d-a26d-94bced60ee78" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.639372 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/certified-operators-lbfd2" podUID="0661ef0e-af9d-41ba-8059-2e878b11b042" containerName="registry-server" probeResult="failure" output="command timed out" Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.639890 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="ca95a4ef-282d-4c26-983f-d7d00f90443c" containerName="galera" probeResult="failure" output="command timed out" Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.639904 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="ca95a4ef-282d-4c26-983f-d7d00f90443c" containerName="galera" probeResult="failure" output="command timed out" Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.762346 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" podUID="a0c610b9-d890-49ee-b321-a3f1efba4b0a" containerName="hostpath-provisioner" probeResult="failure" output="Get \"http://10.217.0.39:9898/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.762405 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="e26e3510-3438-4cb9-8d00-f2d0fdf33ef5" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.171:9090/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.762489 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/prometheus-metric-storage-0" podUID="e26e3510-3438-4cb9-8d00-f2d0fdf33ef5" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.171:9090/-/healthy\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.837488 4934 patch_prober.go:28] interesting pod/nmstate-webhook-f8fb84555-fwjdt container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.86:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.837571 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt" podUID="415e9aa6-5a7a-4d93-924d-0213c9a5ca4d" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.86:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.933400 4934 patch_prober.go:28] interesting pod/logging-loki-gateway-74dbf44c45-4ll7v container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.933464 4934 patch_prober.go:28] interesting pod/logging-loki-gateway-74dbf44c45-4ll7v container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.933517 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" podUID="eac5b26c-d598-4c68-95b5-e2583b456af9" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.54:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.933471 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" podUID="eac5b26c-d598-4c68-95b5-e2583b456af9" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.982630 4934 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.982749 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="642f3213-8063-4ade-800e-4e05ba16d642" containerName="loki-ingester" probeResult="failure" output="Get \"https://10.217.0.55:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.991698 4934 patch_prober.go:28] interesting pod/route-controller-manager-6bd9bdbfbf-x5vvz container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.77:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.991755 4934 patch_prober.go:28] interesting pod/route-controller-manager-6bd9bdbfbf-x5vvz container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.77:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.991941 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" podUID="45f4ca46-4041-4960-8c05-a20bd2c90b90" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.77:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:52 crc kubenswrapper[4934]: I1227 09:07:52.991766 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" podUID="45f4ca46-4041-4960-8c05-a20bd2c90b90" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.77:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:53 crc kubenswrapper[4934]: I1227 09:07:53.065369 4934 patch_prober.go:28] interesting pod/logging-loki-compactor-0 container/loki-compactor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.56:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:53 crc kubenswrapper[4934]: I1227 09:07:53.065419 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-compactor-0" podUID="e6de9198-9266-43e8-b760-e21f0afa885e" containerName="loki-compactor" probeResult="failure" output="Get \"https://10.217.0.56:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:53 crc kubenswrapper[4934]: I1227 09:07:53.133458 4934 patch_prober.go:28] interesting pod/logging-loki-index-gateway-0 container/loki-index-gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:53 crc kubenswrapper[4934]: I1227 09:07:53.133515 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-index-gateway-0" podUID="01f91788-9f82-4508-906b-1e98c9e05c2c" containerName="loki-index-gateway" probeResult="failure" output="Get \"https://10.217.0.57:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:53 crc kubenswrapper[4934]: I1227 09:07:53.293214 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-7rs6w" podUID="ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3" containerName="registry-server" probeResult="failure" output=< Dec 27 09:07:53 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 09:07:53 crc kubenswrapper[4934]: > Dec 27 09:07:53 crc kubenswrapper[4934]: I1227 09:07:53.293838 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/community-operators-sz4sq" podUID="72569b30-4bc6-4f58-b62e-c1a0ef0961d6" containerName="registry-server" probeResult="failure" output=< Dec 27 09:07:53 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 09:07:53 crc kubenswrapper[4934]: > Dec 27 09:07:53 crc kubenswrapper[4934]: I1227 09:07:53.546276 4934 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5jzhr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.63:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:53 crc kubenswrapper[4934]: I1227 09:07:53.546366 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" podUID="a4612dd4-0e0d-4c38-8da5-8bc30dee7c12" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.63:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:53 crc kubenswrapper[4934]: I1227 09:07:53.546307 4934 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5jzhr container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.63:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:53 crc kubenswrapper[4934]: I1227 09:07:53.546424 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" podUID="a4612dd4-0e0d-4c38-8da5-8bc30dee7c12" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.63:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:53 crc kubenswrapper[4934]: I1227 09:07:53.645664 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-handler-kbxfg" podUID="224f9a57-4665-4c0f-865c-ac2267f69b7e" containerName="nmstate-handler" probeResult="failure" output="command timed out" Dec 27 09:07:53 crc kubenswrapper[4934]: I1227 09:07:53.647403 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Dec 27 09:07:53 crc kubenswrapper[4934]: I1227 09:07:53.647609 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 27 09:07:53 crc kubenswrapper[4934]: I1227 09:07:53.690322 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4" podUID="935a4c26-2678-4ead-87ee-1297850d89ff" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.102:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:53 crc kubenswrapper[4934]: I1227 09:07:53.867263 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4" podUID="935a4c26-2678-4ead-87ee-1297850d89ff" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.102:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:53 crc kubenswrapper[4934]: I1227 09:07:53.867351 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx" podUID="fc1c7052-d894-4fa1-b6bb-951cad9e32e0" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.104:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:53 crc kubenswrapper[4934]: I1227 09:07:53.949295 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/glance-operator-controller-manager-64fb555449-tbmpv" podUID="3df60bce-0768-4949-8b8d-3bb28566cd4a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:53 crc kubenswrapper[4934]: I1227 09:07:53.949356 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" podUID="dc4ddef4-e9a5-4934-846e-1d1bcfe9f223" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.103:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.031224 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" podUID="77029090-da44-403b-a20f-d72105281956" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.031223 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" podUID="dc4ddef4-e9a5-4934-846e-1d1bcfe9f223" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.103:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.031330 4934 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Liveness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.031353 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.031386 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/glance-operator-controller-manager-64fb555449-tbmpv" podUID="3df60bce-0768-4949-8b8d-3bb28566cd4a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.031368 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx" podUID="fc1c7052-d894-4fa1-b6bb-951cad9e32e0" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.104:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.113327 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr" podUID="9fbceb4d-e7ae-42c0-8a3f-17807bc9f2ca" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.107:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.113713 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" podUID="77029090-da44-403b-a20f-d72105281956" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.113853 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr" podUID="9fbceb4d-e7ae-42c0-8a3f-17807bc9f2ca" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.107:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.446622 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" podUID="f41746ba-5695-41bc-8078-29ba3d80e3f1" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.446625 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59" podUID="828ebb68-326b-4c32-bc0a-d7f258b45ebd" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.528286 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/neutron-operator-controller-manager-58879495c-d27qh" podUID="89ea143d-f68d-46fc-b15a-6e00a418f65c" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.528469 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" podUID="f41746ba-5695-41bc-8078-29ba3d80e3f1" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.644784 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="40c7cd23-14a7-4b30-9639-84d83ba74fe3" containerName="prometheus" probeResult="failure" output="command timed out" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.644806 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="40c7cd23-14a7-4b30-9639-84d83ba74fe3" containerName="prometheus" probeResult="failure" output="command timed out" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.646236 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.692340 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" podUID="718122b6-c9f4-4d6f-92b5-6862c15f4ee6" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.110:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.692441 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" podUID="1c0960f0-ef55-479a-b375-a09e69254743" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.733482 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" podUID="718122b6-c9f4-4d6f-92b5-6862c15f4ee6" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.110:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.733428 4934 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5fhq7 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.733900 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" podUID="4c71cabc-c13f-4ae8-a704-10efee323e35" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.815290 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59" podUID="828ebb68-326b-4c32-bc0a-d7f258b45ebd" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.856346 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/placement-operator-controller-manager-cc776f956-cj2c7" podUID="7d49c3c9-83c2-4c42-a318-c9d41ddf1bf8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.979502 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/placement-operator-controller-manager-cc776f956-cj2c7" podUID="7d49c3c9-83c2-4c42-a318-c9d41ddf1bf8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:54 crc kubenswrapper[4934]: I1227 09:07:54.979507 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/manila-operator-controller-manager-7cc599445b-dlwjl" podUID="37d56061-eab6-43e5-8dec-f59220da5d0d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.021270 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg" podUID="4f825c5b-90ef-4b56-a759-d7ccbb312bed" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.063334 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/neutron-operator-controller-manager-58879495c-d27qh" podUID="89ea143d-f68d-46fc-b15a-6e00a418f65c" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.063349 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/manila-operator-controller-manager-7cc599445b-dlwjl" podUID="37d56061-eab6-43e5-8dec-f59220da5d0d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.063368 4934 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-2n6w2 container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.063440 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" podUID="964b7818-b04e-4c6d-8ad0-5a108513be73" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.063403 4934 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-2n6w2 container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.063472 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" podUID="964b7818-b04e-4c6d-8ad0-5a108513be73" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.063910 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg" podUID="4f825c5b-90ef-4b56-a759-d7ccbb312bed" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.146703 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" podUID="d090706d-34a7-4c72-9cb9-ad0601db35a6" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.146393 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj" podUID="f2d5e345-1bd9-4313-bdae-7637cb7ed944" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.271380 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" podUID="b5ea792c-b43f-4221-b7bb-aba3421de3d8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.312340 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" podUID="d090706d-34a7-4c72-9cb9-ad0601db35a6" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.312440 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" podUID="b5ea792c-b43f-4221-b7bb-aba3421de3d8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.312469 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" podUID="1c0960f0-ef55-479a-b375-a09e69254743" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.395462 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" podUID="35d6eb8d-3797-432c-a5ea-65f31adf7262" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.397392 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" podUID="59d65d09-c15c-49d9-8ee7-472ffe0dda55" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.481317 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf" podUID="e44e23aa-f466-41cb-b855-d0bdc25cf05b" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.481367 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" podUID="f1cce183-e0a8-4cf1-854e-7a60986f194f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.481401 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" podUID="f1cce183-e0a8-4cf1-854e-7a60986f194f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.481324 4934 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.481443 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.481447 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" podUID="59d65d09-c15c-49d9-8ee7-472ffe0dda55" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.482422 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" podUID="e2db190a-b455-44fc-a43e-2677e5af27b2" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.108:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.482481 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" podUID="e2db190a-b455-44fc-a43e-2677e5af27b2" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.108:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.482695 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj" podUID="f2d5e345-1bd9-4313-bdae-7637cb7ed944" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.483333 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-687f57d79b-5kwg4" podUID="dab3e53b-cbb2-4423-a8e3-432972ecedc3" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.44:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.483479 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf" podUID="e44e23aa-f466-41cb-b855-d0bdc25cf05b" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.481920 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" podUID="35d6eb8d-3797-432c-a5ea-65f31adf7262" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.751261 4934 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-lnknm container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.751315 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" podUID="4575d285-879f-4cc3-a9b9-406dc96e852d" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.751401 4934 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-lnknm container/operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.751434 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" podUID="4575d285-879f-4cc3-a9b9-406dc96e852d" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.751787 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" podUID="a0c610b9-d890-49ee-b321-a3f1efba4b0a" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.876371 4934 patch_prober.go:28] interesting pod/perses-operator-5bf474d74f-tx6z5 container/perses-operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.13:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.876443 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" podUID="f97e8d86-1e9e-4c5a-a164-2600193c576b" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.13:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.985231 4934 patch_prober.go:28] interesting pod/console-74cc7c486-2dpkp container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.138:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:55 crc kubenswrapper[4934]: I1227 09:07:55.985296 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-74cc7c486-2dpkp" podUID="e0978ad2-1582-49d4-b239-ff929dafb117" containerName="console" probeResult="failure" output="Get \"https://10.217.0.138:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:56 crc kubenswrapper[4934]: I1227 09:07:56.054295 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc" podUID="b461380d-6573-4898-9bc8-c82ceba822d8" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.101:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:56 crc kubenswrapper[4934]: I1227 09:07:56.239946 4934 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5fhq7 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Dec 27 09:07:56 crc kubenswrapper[4934]: I1227 09:07:56.240007 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" podUID="4c71cabc-c13f-4ae8-a704-10efee323e35" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Dec 27 09:07:56 crc kubenswrapper[4934]: I1227 09:07:56.391374 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="4e39152f-3592-4907-9b06-d13074aeb3c4" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.5:8080/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:56 crc kubenswrapper[4934]: I1227 09:07:56.391202 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="4e39152f-3592-4907-9b06-d13074aeb3c4" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.5:8081/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:56 crc kubenswrapper[4934]: I1227 09:07:56.754154 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" podUID="a0c610b9-d890-49ee-b321-a3f1efba4b0a" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 09:07:56 crc kubenswrapper[4934]: I1227 09:07:56.923690 4934 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-hzmws container/oauth-apiserver namespace/openshift-oauth-apiserver: Liveness probe status=failure output="Get \"https://10.217.0.28:8443/livez?exclude=etcd\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:56 crc kubenswrapper[4934]: I1227 09:07:56.923774 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" podUID="145ee02a-c7aa-4990-a288-e6f69f8ab030" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.28:8443/livez?exclude=etcd\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:56 crc kubenswrapper[4934]: I1227 09:07:56.924228 4934 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-hzmws container/oauth-apiserver namespace/openshift-oauth-apiserver: Readiness probe status=failure output="Get \"https://10.217.0.28:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:56 crc kubenswrapper[4934]: I1227 09:07:56.924252 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" podUID="145ee02a-c7aa-4990-a288-e6f69f8ab030" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.28:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.064347 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" podUID="fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.93:8080/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.426317 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" podUID="df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.94:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.426490 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" podUID="df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.94:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.452323 4934 patch_prober.go:28] interesting pod/thanos-querier-8467887986-svb7h container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.71:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.452385 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-8467887986-svb7h" podUID="a07fdfc4-584a-4a39-b87f-ee78f9cac687" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.71:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.452420 4934 patch_prober.go:28] interesting pod/thanos-querier-8467887986-svb7h container/kube-rbac-proxy-web namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.71:9091/-/healthy\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.452492 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/thanos-querier-8467887986-svb7h" podUID="a07fdfc4-584a-4a39-b87f-ee78f9cac687" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.71:9091/-/healthy\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.503070 4934 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:6443/livez?exclude=etcd\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.503136 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez?exclude=etcd\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.503465 4934 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.503484 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.560137 4934 patch_prober.go:28] interesting pod/logging-loki-gateway-74dbf44c45-qd69g container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.560233 4934 patch_prober.go:28] interesting pod/logging-loki-gateway-74dbf44c45-qd69g container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.560269 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" podUID="47cbee2e-c325-4a8d-a26d-94bced60ee78" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.560304 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" podUID="47cbee2e-c325-4a8d-a26d-94bced60ee78" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.53:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.641793 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="40c7cd23-14a7-4b30-9639-84d83ba74fe3" containerName="prometheus" probeResult="failure" output="command timed out" Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.671465 4934 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-dwv8c container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.671520 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-dwv8c" podUID="901df0a0-1dae-41aa-8d29-0981524d9312" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.743545 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="e26e3510-3438-4cb9-8d00-f2d0fdf33ef5" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.171:9090/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.743903 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/prometheus-metric-storage-0" podUID="e26e3510-3438-4cb9-8d00-f2d0fdf33ef5" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.171:9090/-/healthy\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.744599 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.873274 4934 patch_prober.go:28] interesting pod/downloads-7954f5f757-8msj4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.873338 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8msj4" podUID="a0d34185-bf08-456c-a5fc-3f0852d35a91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.873402 4934 patch_prober.go:28] interesting pod/downloads-7954f5f757-8msj4 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.873416 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-8msj4" podUID="a0d34185-bf08-456c-a5fc-3f0852d35a91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.887130 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" event={"ID":"00799e47-9b94-49bf-8bd1-d6bb5036285a","Type":"ContainerDied","Data":"ed64f69a04114620942ed0f14a24739b1d58dd40935c3a97801993860c86d312"} Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.894050 4934 scope.go:117] "RemoveContainer" containerID="ed64f69a04114620942ed0f14a24739b1d58dd40935c3a97801993860c86d312" Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.895448 4934 generic.go:334] "Generic (PLEG): container finished" podID="00799e47-9b94-49bf-8bd1-d6bb5036285a" containerID="ed64f69a04114620942ed0f14a24739b1d58dd40935c3a97801993860c86d312" exitCode=1 Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.933516 4934 patch_prober.go:28] interesting pod/logging-loki-gateway-74dbf44c45-4ll7v container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.933550 4934 patch_prober.go:28] interesting pod/logging-loki-gateway-74dbf44c45-4ll7v container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.933576 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" podUID="eac5b26c-d598-4c68-95b5-e2583b456af9" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.54:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:57 crc kubenswrapper[4934]: I1227 09:07:57.933569 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" podUID="eac5b26c-d598-4c68-95b5-e2583b456af9" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.007254 4934 patch_prober.go:28] interesting pod/console-operator-58897d9998-jckgt container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.007352 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-jckgt" podUID="753c599b-a401-481e-a5b3-e2dd6ed2c4e7" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.007270 4934 patch_prober.go:28] interesting pod/console-operator-58897d9998-jckgt container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.008201 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-jckgt" podUID="753c599b-a401-481e-a5b3-e2dd6ed2c4e7" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.175902 4934 patch_prober.go:28] interesting pod/image-registry-66df7c8f76-vkfr4 container/registry namespace/openshift-image-registry: Readiness probe status=failure output="Get \"https://10.217.0.78:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.176271 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" podUID="21dab295-f1e7-4fee-a281-251122757a57" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.78:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.179125 4934 patch_prober.go:28] interesting pod/image-registry-66df7c8f76-vkfr4 container/registry namespace/openshift-image-registry: Liveness probe status=failure output="Get \"https://10.217.0.78:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.179188 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-image-registry/image-registry-66df7c8f76-vkfr4" podUID="21dab295-f1e7-4fee-a281-251122757a57" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.78:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.334331 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5bsq9" podUID="0134c36f-c43b-4830-848a-3a8690957ee1" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.95:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.375289 4934 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-5cc9p container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.375334 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-7784b6fcf-5bsq9" podUID="0134c36f-c43b-4830-848a-3a8690957ee1" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.95:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.375385 4934 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-5cc9p container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.375362 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" podUID="af023fee-eae1-427a-8d70-487588422316" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.375459 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" podUID="af023fee-eae1-427a-8d70-487588422316" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.437282 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/controller-5bddd4b946-7c5ls" podUID="a8825499-558f-4b31-b09b-0dcc92662f65" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.96:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.479335 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/controller-5bddd4b946-7c5ls" podUID="a8825499-558f-4b31-b09b-0dcc92662f65" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.96:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.643226 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-hrs9g" podUID="73bc5c15-5c9d-49ae-bcd9-853df82fabbf" containerName="registry-server" probeResult="failure" output="command timed out" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.643226 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.643323 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.645021 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-central-agent" containerStatusID={"Type":"cri-o","ID":"9995c35f73f77d890191ce4b613fd43a168910498934a2d034fa50716244fb04"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-central-agent failed liveness probe, will be restarted" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.645133 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="ceilometer-central-agent" containerID="cri-o://9995c35f73f77d890191ce4b613fd43a168910498934a2d034fa50716244fb04" gracePeriod=30 Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.674442 4934 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-r6jld container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.674534 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" podUID="4ea539b6-ffe0-4393-af8d-62579266697c" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.674442 4934 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-r6jld container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.674665 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" podUID="4ea539b6-ffe0-4393-af8d-62579266697c" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.762530 4934 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-t6q48 container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.763124 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" podUID="873d8371-2f45-409a-b18d-f4c7dbdabeea" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.762570 4934 patch_prober.go:28] interesting pod/router-default-5444994796-cpts5 container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.763206 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-cpts5" podUID="ac1de940-6952-466a-be96-52c43b1a90a2" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.762667 4934 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-t6q48 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.763242 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" podUID="873d8371-2f45-409a-b18d-f4c7dbdabeea" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.762705 4934 patch_prober.go:28] interesting pod/router-default-5444994796-cpts5 container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.763275 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-cpts5" podUID="ac1de940-6952-466a-be96-52c43b1a90a2" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.953521 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-v2lll" podUID="c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.998874 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-v2lll" podUID="c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:58 crc kubenswrapper[4934]: I1227 09:07:58.999267 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-v2lll" podUID="c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:59 crc kubenswrapper[4934]: I1227 09:07:59.082311 4934 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-c2tqv container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:59 crc kubenswrapper[4934]: I1227 09:07:59.082402 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-c2tqv" podUID="85b3562e-4570-4b9b-baa5-a20fe55f13ca" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:59 crc kubenswrapper[4934]: I1227 09:07:59.082502 4934 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-c2tqv container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:59 crc kubenswrapper[4934]: I1227 09:07:59.082583 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-c2tqv" podUID="85b3562e-4570-4b9b-baa5-a20fe55f13ca" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:59 crc kubenswrapper[4934]: I1227 09:07:59.082650 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-5s9zf" podUID="a0c610b9-d890-49ee-b321-a3f1efba4b0a" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 09:07:59 crc kubenswrapper[4934]: I1227 09:07:59.259232 4934 patch_prober.go:28] interesting pod/oauth-openshift-fd9565bb5-mkr88 container/oauth-openshift namespace/openshift-authentication: Liveness probe status=failure output="Get \"https://10.217.0.72:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:59 crc kubenswrapper[4934]: I1227 09:07:59.259288 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" podUID="772a3e80-411a-4c15-9e1c-b702392e13af" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.72:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:59 crc kubenswrapper[4934]: I1227 09:07:59.259232 4934 patch_prober.go:28] interesting pod/oauth-openshift-fd9565bb5-mkr88 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.72:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:07:59 crc kubenswrapper[4934]: I1227 09:07:59.259347 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" podUID="772a3e80-411a-4c15-9e1c-b702392e13af" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.72:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:59 crc kubenswrapper[4934]: I1227 09:07:59.260012 4934 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5fhq7 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Dec 27 09:07:59 crc kubenswrapper[4934]: I1227 09:07:59.260104 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" podUID="4c71cabc-c13f-4ae8-a704-10efee323e35" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Dec 27 09:07:59 crc kubenswrapper[4934]: I1227 09:07:59.468648 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:07:59 crc kubenswrapper[4934]: E1227 09:07:59.491666 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:07:59 crc kubenswrapper[4934]: I1227 09:07:59.640504 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-northd-0" podUID="727f882e-fa63-46e9-9cd5-9781e472f2ef" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 27 09:07:59 crc kubenswrapper[4934]: I1227 09:07:59.640879 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="727f882e-fa63-46e9-9cd5-9781e472f2ef" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 27 09:07:59 crc kubenswrapper[4934]: I1227 09:07:59.641412 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="40c7cd23-14a7-4b30-9639-84d83ba74fe3" containerName="prometheus" probeResult="failure" output="command timed out" Dec 27 09:07:59 crc kubenswrapper[4934]: I1227 09:07:59.955224 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-4jwwj" podUID="ea01c7df-c5aa-4c86-800b-b6aebafeb8fd" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:07:59 crc kubenswrapper[4934]: I1227 09:07:59.955260 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-4jwwj" podUID="ea01c7df-c5aa-4c86-800b-b6aebafeb8fd" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:00 crc kubenswrapper[4934]: I1227 09:08:00.400165 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 09:08:00 crc kubenswrapper[4934]: I1227 09:08:00.400423 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 09:08:00 crc kubenswrapper[4934]: I1227 09:08:00.640158 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="40c7cd23-14a7-4b30-9639-84d83ba74fe3" containerName="prometheus" probeResult="failure" output="command timed out" Dec 27 09:08:00 crc kubenswrapper[4934]: I1227 09:08:00.740850 4934 patch_prober.go:28] interesting pod/metrics-server-76d5d9b996-jzgqz container/metrics-server namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.74:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:00 crc kubenswrapper[4934]: I1227 09:08:00.741523 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/metrics-server-76d5d9b996-jzgqz" podUID="169cb9e6-0d57-425d-a51d-a3b0e1556db6" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.74:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:00 crc kubenswrapper[4934]: I1227 09:08:00.745894 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="e26e3510-3438-4cb9-8d00-f2d0fdf33ef5" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.171:9090/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.129730 4934 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.129810 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.263769 4934 patch_prober.go:28] interesting pod/monitoring-plugin-7d4f6975b-p9h8q container/monitoring-plugin namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.75:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.263854 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/monitoring-plugin-7d4f6975b-p9h8q" podUID="b78888a7-451e-43a3-86ea-b13c035afab1" containerName="monitoring-plugin" probeResult="failure" output="Get \"https://10.217.0.75:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.480199 4934 patch_prober.go:28] interesting pod/controller-manager-7967777f79-kzj4d container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.83:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.480269 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" podUID="2a688ece-0e00-4241-b7dd-9dde25f0eaf1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.83:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.480364 4934 patch_prober.go:28] interesting pod/controller-manager-7967777f79-kzj4d container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.83:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.480380 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" podUID="2a688ece-0e00-4241-b7dd-9dde25f0eaf1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.83:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.500226 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.500333 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.502440 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="controller-manager" containerStatusID={"Type":"cri-o","ID":"fd15bc9e9543c8afdb985cb0426e5686fff142c76fd4188ddea84d8885e949cc"} pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" containerMessage="Container controller-manager failed liveness probe, will be restarted" Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.502673 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" podUID="2a688ece-0e00-4241-b7dd-9dde25f0eaf1" containerName="controller-manager" containerID="cri-o://fd15bc9e9543c8afdb985cb0426e5686fff142c76fd4188ddea84d8885e949cc" gracePeriod=30 Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.640424 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="d03cf582-2192-4a10-b21e-10befebde419" containerName="galera" probeResult="failure" output="command timed out" Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.642826 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="d03cf582-2192-4a10-b21e-10befebde419" containerName="galera" probeResult="failure" output="command timed out" Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.642892 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-index-v7q5s" podUID="a41afe73-59e9-4cbe-b2e7-1f755767daf5" containerName="registry-server" probeResult="failure" output="command timed out" Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.643869 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-operators-qtq22" podUID="2ee60051-8eec-4da3-ae78-b409c5cdf9f3" containerName="registry-server" probeResult="failure" output="command timed out" Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.645800 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-operators-qtq22" podUID="2ee60051-8eec-4da3-ae78-b409c5cdf9f3" containerName="registry-server" probeResult="failure" output="command timed out" Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.649608 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-index-v7q5s" podUID="a41afe73-59e9-4cbe-b2e7-1f755767daf5" containerName="registry-server" probeResult="failure" output="command timed out" Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.805764 4934 patch_prober.go:28] interesting pod/logging-loki-distributor-5f678c8dd6-nqqn7 container/loki-distributor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.50:3101/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.806107 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" podUID="5864b138-b8ef-405e-9ed9-be39f8e13e8d" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.50:3101/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.806198 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.952363 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" event={"ID":"00799e47-9b94-49bf-8bd1-d6bb5036285a","Type":"ContainerStarted","Data":"73727fb66c1e09152eba8a4df3ec72bd09c3c790d205c3451f0b982c8fd1d2e5"} Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.952441 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.954114 4934 patch_prober.go:28] interesting pod/apiserver-76f77b778f-8s8bp container/openshift-apiserver namespace/openshift-apiserver: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/readyz?exclude=etcd&exclude=etcd-readiness\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:01 crc kubenswrapper[4934]: I1227 09:08:01.954161 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-apiserver/apiserver-76f77b778f-8s8bp" podUID="d812eb7a-ce30-4759-a558-fff3738068b6" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.25:8443/readyz?exclude=etcd&exclude=etcd-readiness\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.068518 4934 patch_prober.go:28] interesting pod/logging-loki-query-frontend-69d9546745-4b9rt container/loki-query-frontend namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.52:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.068600 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" podUID="b9e3cc00-7197-4c75-a732-136e65d893c8" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.52:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.068711 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.239629 4934 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5fhq7 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.239694 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" podUID="4c71cabc-c13f-4ae8-a704-10efee323e35" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.268364 4934 patch_prober.go:28] interesting pod/logging-loki-querier-76788598db-6wrwh container/loki-querier namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.51:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.268449 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" podUID="ae25e4b9-7a2a-41ff-9944-d6c45603cbec" containerName="loki-querier" probeResult="failure" output="Get \"https://10.217.0.51:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.268543 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.451606 4934 patch_prober.go:28] interesting pod/thanos-querier-8467887986-svb7h container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.71:9091/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.451662 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-8467887986-svb7h" podUID="a07fdfc4-584a-4a39-b87f-ee78f9cac687" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.71:9091/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.560047 4934 patch_prober.go:28] interesting pod/logging-loki-gateway-74dbf44c45-qd69g container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.560063 4934 patch_prober.go:28] interesting pod/logging-loki-gateway-74dbf44c45-qd69g container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.560165 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" podUID="47cbee2e-c325-4a8d-a26d-94bced60ee78" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.560192 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" podUID="47cbee2e-c325-4a8d-a26d-94bced60ee78" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.53:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.639799 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="ca95a4ef-282d-4c26-983f-d7d00f90443c" containerName="galera" probeResult="failure" output="command timed out" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.639922 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.641137 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="ca95a4ef-282d-4c26-983f-d7d00f90443c" containerName="galera" probeResult="failure" output="command timed out" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.641245 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.644445 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/certified-operators-lbfd2" podUID="0661ef0e-af9d-41ba-8059-2e878b11b042" containerName="registry-server" probeResult="failure" output="command timed out" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.644500 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/certified-operators-lbfd2" podUID="0661ef0e-af9d-41ba-8059-2e878b11b042" containerName="registry-server" probeResult="failure" output="command timed out" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.806605 4934 patch_prober.go:28] interesting pod/logging-loki-distributor-5f678c8dd6-nqqn7 container/loki-distributor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.50:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.806705 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" podUID="5864b138-b8ef-405e-9ed9-be39f8e13e8d" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.50:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.837168 4934 patch_prober.go:28] interesting pod/nmstate-webhook-f8fb84555-fwjdt container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.86:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.837584 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt" podUID="415e9aa6-5a7a-4d93-924d-0213c9a5ca4d" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.86:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.838036 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.932739 4934 patch_prober.go:28] interesting pod/logging-loki-gateway-74dbf44c45-4ll7v container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.932793 4934 patch_prober.go:28] interesting pod/logging-loki-gateway-74dbf44c45-4ll7v container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.932803 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" podUID="eac5b26c-d598-4c68-95b5-e2583b456af9" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.932850 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" podUID="eac5b26c-d598-4c68-95b5-e2583b456af9" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.54:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.960611 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="galera" containerStatusID={"Type":"cri-o","ID":"e862df5ec5f94dc038eceb2b1533dcf21f0a38c95f61e193874266056fd27165"} pod="openstack/openstack-cell1-galera-0" containerMessage="Container galera failed liveness probe, will be restarted" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.982786 4934 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.983066 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="642f3213-8063-4ade-800e-4e05ba16d642" containerName="loki-ingester" probeResult="failure" output="Get \"https://10.217.0.55:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.983156 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.990887 4934 patch_prober.go:28] interesting pod/route-controller-manager-6bd9bdbfbf-x5vvz container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.77:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.990946 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" podUID="45f4ca46-4041-4960-8c05-a20bd2c90b90" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.77:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.991002 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.991389 4934 patch_prober.go:28] interesting pod/route-controller-manager-6bd9bdbfbf-x5vvz container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.77:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.991469 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" podUID="45f4ca46-4041-4960-8c05-a20bd2c90b90" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.77:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.992176 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="route-controller-manager" containerStatusID={"Type":"cri-o","ID":"5f66e5c0191c6ce691d9d5526a2dabacd95623f4549ca175a37bdd114a0a0a6d"} pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" containerMessage="Container route-controller-manager failed liveness probe, will be restarted" Dec 27 09:08:02 crc kubenswrapper[4934]: I1227 09:08:02.992215 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" podUID="45f4ca46-4041-4960-8c05-a20bd2c90b90" containerName="route-controller-manager" containerID="cri-o://5f66e5c0191c6ce691d9d5526a2dabacd95623f4549ca175a37bdd114a0a0a6d" gracePeriod=30 Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.065456 4934 patch_prober.go:28] interesting pod/logging-loki-compactor-0 container/loki-compactor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.56:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.065541 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-compactor-0" podUID="e6de9198-9266-43e8-b760-e21f0afa885e" containerName="loki-compactor" probeResult="failure" output="Get \"https://10.217.0.56:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.065611 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.069043 4934 patch_prober.go:28] interesting pod/logging-loki-query-frontend-69d9546745-4b9rt container/loki-query-frontend namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.52:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.069092 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" podUID="b9e3cc00-7197-4c75-a732-136e65d893c8" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.52:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.133477 4934 patch_prober.go:28] interesting pod/logging-loki-index-gateway-0 container/loki-index-gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.133528 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-index-gateway-0" podUID="01f91788-9f82-4508-906b-1e98c9e05c2c" containerName="loki-index-gateway" probeResult="failure" output="Get \"https://10.217.0.57:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.133594 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.202058 4934 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.214511 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.215167 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-7rs6w" podUID="ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3" containerName="registry-server" probeResult="failure" output=< Dec 27 09:08:03 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 09:08:03 crc kubenswrapper[4934]: > Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.215387 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/community-operators-sz4sq" podUID="72569b30-4bc6-4f58-b62e-c1a0ef0961d6" containerName="registry-server" probeResult="failure" output=< Dec 27 09:08:03 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 09:08:03 crc kubenswrapper[4934]: > Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.215430 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-sz4sq" podUID="72569b30-4bc6-4f58-b62e-c1a0ef0961d6" containerName="registry-server" probeResult="failure" output=< Dec 27 09:08:03 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 09:08:03 crc kubenswrapper[4934]: > Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.270357 4934 patch_prober.go:28] interesting pod/logging-loki-querier-76788598db-6wrwh container/loki-querier namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.51:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.270486 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" podUID="ae25e4b9-7a2a-41ff-9944-d6c45603cbec" containerName="loki-querier" probeResult="failure" output="Get \"https://10.217.0.51:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.292792 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-marketplace-7rs6w" podUID="ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3" containerName="registry-server" probeResult="failure" output=< Dec 27 09:08:03 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 09:08:03 crc kubenswrapper[4934]: > Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.510207 4934 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5jzhr container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.63:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.510486 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" podUID="a4612dd4-0e0d-4c38-8da5-8bc30dee7c12" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.63:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.510532 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.511962 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="marketplace-operator" containerStatusID={"Type":"cri-o","ID":"67e9d2fa2404b9f860a2bbc81d950b59caee85958a0d2750c45a0fcbd374a629"} pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" containerMessage="Container marketplace-operator failed liveness probe, will be restarted" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.512094 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" podUID="a4612dd4-0e0d-4c38-8da5-8bc30dee7c12" containerName="marketplace-operator" containerID="cri-o://67e9d2fa2404b9f860a2bbc81d950b59caee85958a0d2750c45a0fcbd374a629" gracePeriod=30 Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.551872 4934 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5jzhr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.63:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.551940 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" podUID="a4612dd4-0e0d-4c38-8da5-8bc30dee7c12" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.63:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.552021 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.639192 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="ca95a4ef-282d-4c26-983f-d7d00f90443c" containerName="galera" probeResult="failure" output="command timed out" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.733251 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx" podUID="fc1c7052-d894-4fa1-b6bb-951cad9e32e0" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.104:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.733254 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4" podUID="935a4c26-2678-4ead-87ee-1297850d89ff" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.102:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.733602 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" podUID="dc4ddef4-e9a5-4934-846e-1d1bcfe9f223" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.103:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.734313 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.734378 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.736121 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.774248 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/glance-operator-controller-manager-64fb555449-tbmpv" podUID="3df60bce-0768-4949-8b8d-3bb28566cd4a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.774343 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-64fb555449-tbmpv" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.815514 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" podUID="77029090-da44-403b-a20f-d72105281956" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.815635 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.839292 4934 patch_prober.go:28] interesting pod/nmstate-webhook-f8fb84555-fwjdt container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.86:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.839352 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt" podUID="415e9aa6-5a7a-4d93-924d-0213c9a5ca4d" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.86:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.855863 4934 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Liveness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.856061 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.856145 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.857921 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-scheduler" containerStatusID={"Type":"cri-o","ID":"80370fd8607ddb05cccbe7383b94c5a4c1a874299bbbc7f665805ae5edf73ef9"} pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" containerMessage="Container kube-scheduler failed liveness probe, will be restarted" Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.858003 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" containerID="cri-o://80370fd8607ddb05cccbe7383b94c5a4c1a874299bbbc7f665805ae5edf73ef9" gracePeriod=30 Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.908346 4934 trace.go:236] Trace[1065527890]: "Calculate volume metrics of ovndbcluster-sb-etc-ovn for pod openstack/ovsdbserver-sb-0" (27-Dec-2025 09:08:01.193) (total time: 2713ms): Dec 27 09:08:03 crc kubenswrapper[4934]: Trace[1065527890]: [2.713840785s] [2.713840785s] END Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.908349 4934 trace.go:236] Trace[58859840]: "Calculate volume metrics of storage for pod minio-dev/minio" (27-Dec-2025 09:07:59.702) (total time: 4204ms): Dec 27 09:08:03 crc kubenswrapper[4934]: Trace[58859840]: [4.20466445s] [4.20466445s] END Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.986644 4934 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:03 crc kubenswrapper[4934]: I1227 09:08:03.986697 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="642f3213-8063-4ade-800e-4e05ba16d642" containerName="loki-ingester" probeResult="failure" output="Get \"https://10.217.0.55:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.066274 4934 patch_prober.go:28] interesting pod/logging-loki-compactor-0 container/loki-compactor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.56:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.066597 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-compactor-0" podUID="e6de9198-9266-43e8-b760-e21f0afa885e" containerName="loki-compactor" probeResult="failure" output="Get \"https://10.217.0.56:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.079807 4934 trace.go:236] Trace[2053784379]: "Calculate volume metrics of storage for pod openshift-logging/logging-loki-compactor-0" (27-Dec-2025 09:07:52.886) (total time: 11193ms): Dec 27 09:08:04 crc kubenswrapper[4934]: Trace[2053784379]: [11.193341181s] [11.193341181s] END Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.134544 4934 patch_prober.go:28] interesting pod/logging-loki-index-gateway-0 container/loki-index-gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.134613 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-index-gateway-0" podUID="01f91788-9f82-4508-906b-1e98c9e05c2c" containerName="loki-index-gateway" probeResult="failure" output="Get \"https://10.217.0.57:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.174254 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" podUID="718122b6-c9f4-4d6f-92b5-6862c15f4ee6" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.110:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.174371 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.182507 4934 trace.go:236] Trace[1954623194]: "Calculate volume metrics of ovndbcluster-nb-etc-ovn for pod openstack/ovsdbserver-nb-0" (27-Dec-2025 09:08:00.807) (total time: 3374ms): Dec 27 09:08:04 crc kubenswrapper[4934]: Trace[1954623194]: [3.374936693s] [3.374936693s] END Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.182552 4934 trace.go:236] Trace[1444450247]: "Calculate volume metrics of glance for pod openstack/glance-default-internal-api-0" (27-Dec-2025 09:07:59.882) (total time: 4300ms): Dec 27 09:08:04 crc kubenswrapper[4934]: Trace[1444450247]: [4.300345087s] [4.300345087s] END Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.182613 4934 trace.go:236] Trace[1818105531]: "Calculate volume metrics of mysql-db for pod openstack/openstack-cell1-galera-0" (27-Dec-2025 09:07:54.802) (total time: 9380ms): Dec 27 09:08:04 crc kubenswrapper[4934]: Trace[1818105531]: [9.380036721s] [9.380036721s] END Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.182765 4934 trace.go:236] Trace[2026698856]: "Calculate volume metrics of swift for pod openstack/swift-storage-0" (27-Dec-2025 09:08:00.361) (total time: 3821ms): Dec 27 09:08:04 crc kubenswrapper[4934]: Trace[2026698856]: [3.821061649s] [3.821061649s] END Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.216346 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59" podUID="828ebb68-326b-4c32-bc0a-d7f258b45ebd" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.216437 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.216898 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" podUID="f41746ba-5695-41bc-8078-29ba3d80e3f1" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.259401 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/neutron-operator-controller-manager-58879495c-d27qh" podUID="89ea143d-f68d-46fc-b15a-6e00a418f65c" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.259743 4934 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": dial tcp 192.168.126.11:10259: connect: connection refused" start-of-body= Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.259803 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": dial tcp 192.168.126.11:10259: connect: connection refused" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.260104 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.341277 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" podUID="1c0960f0-ef55-479a-b375-a09e69254743" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.341398 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.341396 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" podUID="d090706d-34a7-4c72-9cb9-ad0601db35a6" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.341503 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-69977bdf55-p5jgx" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.341609 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.382518 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" podUID="59d65d09-c15c-49d9-8ee7-472ffe0dda55" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.382710 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.387457 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-669b58f65-t89p4" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.436300 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg" podUID="4f825c5b-90ef-4b56-a759-d7ccbb312bed" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.436437 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.515695 4934 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-2n6w2 container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.515755 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" podUID="964b7818-b04e-4c6d-8ad0-5a108513be73" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.515837 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.515929 4934 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-2n6w2 container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.515980 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" podUID="964b7818-b04e-4c6d-8ad0-5a108513be73" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.516025 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.517101 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="prometheus-operator-admission-webhook" containerStatusID={"Type":"cri-o","ID":"bf4828a80aa2309175c3712a6b88e7388112737fa373bc6e393056a56a3ad929"} pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" containerMessage="Container prometheus-operator-admission-webhook failed liveness probe, will be restarted" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.517141 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" podUID="964b7818-b04e-4c6d-8ad0-5a108513be73" containerName="prometheus-operator-admission-webhook" containerID="cri-o://bf4828a80aa2309175c3712a6b88e7388112737fa373bc6e393056a56a3ad929" gracePeriod=30 Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.639523 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="727f882e-fa63-46e9-9cd5-9781e472f2ef" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.640409 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-northd-0" podUID="727f882e-fa63-46e9-9cd5-9781e472f2ef" containerName="ovn-northd" probeResult="failure" output="command timed out" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.641049 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="40c7cd23-14a7-4b30-9639-84d83ba74fe3" containerName="prometheus" probeResult="failure" output="command timed out" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.641437 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="ca95a4ef-282d-4c26-983f-d7d00f90443c" containerName="galera" probeResult="failure" output="command timed out" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.645051 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="40c7cd23-14a7-4b30-9639-84d83ba74fe3" containerName="prometheus" probeResult="failure" output="command timed out" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.661253 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" podUID="b5ea792c-b43f-4221-b7bb-aba3421de3d8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.661383 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.662196 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" podUID="f1cce183-e0a8-4cf1-854e-7a60986f194f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.662284 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.741617 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" podUID="35d6eb8d-3797-432c-a5ea-65f31adf7262" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.741756 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.783282 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" podUID="dc4ddef4-e9a5-4934-846e-1d1bcfe9f223" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.103:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.865513 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf" podUID="e44e23aa-f466-41cb-b855-d0bdc25cf05b" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.865993 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.866059 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.907266 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/glance-operator-controller-manager-64fb555449-tbmpv" podUID="3df60bce-0768-4949-8b8d-3bb28566cd4a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.105:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:04 crc kubenswrapper[4934]: I1227 09:08:04.907266 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" podUID="77029090-da44-403b-a20f-d72105281956" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.215349 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" podUID="718122b6-c9f4-4d6f-92b5-6862c15f4ee6" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.110:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.259260 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59" podUID="828ebb68-326b-4c32-bc0a-d7f258b45ebd" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.259753 4934 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5fhq7 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.259876 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" podUID="4c71cabc-c13f-4ae8-a704-10efee323e35" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.290311 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-687f57d79b-5kwg4" podUID="dab3e53b-cbb2-4423-a8e3-432972ecedc3" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.44:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.424383 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" podUID="1c0960f0-ef55-479a-b375-a09e69254743" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.424841 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" podUID="d090706d-34a7-4c72-9cb9-ad0601db35a6" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.506291 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" podUID="59d65d09-c15c-49d9-8ee7-472ffe0dda55" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.506264 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg" podUID="4f825c5b-90ef-4b56-a759-d7ccbb312bed" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.517001 4934 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-2n6w2 container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.517095 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" podUID="964b7818-b04e-4c6d-8ad0-5a108513be73" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.788279 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" podUID="f1cce183-e0a8-4cf1-854e-7a60986f194f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.788333 4934 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-lnknm container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.788433 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" podUID="4575d285-879f-4cc3-a9b9-406dc96e852d" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.788541 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.830304 4934 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-lnknm container/operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.830592 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" podUID="4575d285-879f-4cc3-a9b9-406dc96e852d" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.830339 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" podUID="35d6eb8d-3797-432c-a5ea-65f31adf7262" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.830662 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.959324 4934 patch_prober.go:28] interesting pod/perses-operator-5bf474d74f-tx6z5 container/perses-operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.13:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.959399 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf" podUID="e44e23aa-f466-41cb-b855-d0bdc25cf05b" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.959613 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" podUID="f97e8d86-1e9e-4c5a-a164-2600193c576b" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.13:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.959437 4934 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": EOF" start-of-body= Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.959902 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.959470 4934 patch_prober.go:28] interesting pod/perses-operator-5bf474d74f-tx6z5 container/perses-operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.13:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.959916 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": EOF" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.959954 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" podUID="f97e8d86-1e9e-4c5a-a164-2600193c576b" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.13:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.987068 4934 patch_prober.go:28] interesting pod/console-74cc7c486-2dpkp container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.138:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.987303 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-74cc7c486-2dpkp" podUID="e0978ad2-1582-49d4-b239-ff929dafb117" containerName="console" probeResult="failure" output="Get \"https://10.217.0.138:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.987459 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.996814 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="operator" containerStatusID={"Type":"cri-o","ID":"0e9c0d7603353e305c0f51f83e3c3c23c80d0364f79fc2f47a4a05801de2061b"} pod="openshift-operators/observability-operator-59bdc8b94-lnknm" containerMessage="Container operator failed liveness probe, will be restarted" Dec 27 09:08:05 crc kubenswrapper[4934]: I1227 09:08:05.997120 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" podUID="4575d285-879f-4cc3-a9b9-406dc96e852d" containerName="operator" containerID="cri-o://0e9c0d7603353e305c0f51f83e3c3c23c80d0364f79fc2f47a4a05801de2061b" gracePeriod=30 Dec 27 09:08:06 crc kubenswrapper[4934]: I1227 09:08:06.095359 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc" podUID="b461380d-6573-4898-9bc8-c82ceba822d8" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.101:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:06 crc kubenswrapper[4934]: I1227 09:08:06.095423 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc" podUID="b461380d-6573-4898-9bc8-c82ceba822d8" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.101:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:06 crc kubenswrapper[4934]: I1227 09:08:06.095547 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc" Dec 27 09:08:06 crc kubenswrapper[4934]: I1227 09:08:06.354259 4934 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 27 09:08:06 crc kubenswrapper[4934]: I1227 09:08:06.354322 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 27 09:08:06 crc kubenswrapper[4934]: I1227 09:08:06.391092 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="4e39152f-3592-4907-9b06-d13074aeb3c4" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.5:8080/livez\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:06 crc kubenswrapper[4934]: I1227 09:08:06.391259 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/kube-state-metrics-0" Dec 27 09:08:06 crc kubenswrapper[4934]: I1227 09:08:06.392533 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-state-metrics" containerStatusID={"Type":"cri-o","ID":"2437eadde08e055b94e28a357e0995921db59807bc010af89911716c3aabd6e0"} pod="openstack/kube-state-metrics-0" containerMessage="Container kube-state-metrics failed liveness probe, will be restarted" Dec 27 09:08:06 crc kubenswrapper[4934]: I1227 09:08:06.392652 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="4e39152f-3592-4907-9b06-d13074aeb3c4" containerName="kube-state-metrics" containerID="cri-o://2437eadde08e055b94e28a357e0995921db59807bc010af89911716c3aabd6e0" gracePeriod=30 Dec 27 09:08:06 crc kubenswrapper[4934]: I1227 09:08:06.831428 4934 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-lnknm container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:06 crc kubenswrapper[4934]: I1227 09:08:06.831781 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" podUID="4575d285-879f-4cc3-a9b9-406dc96e852d" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.16:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:06 crc kubenswrapper[4934]: I1227 09:08:06.924434 4934 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-hzmws container/oauth-apiserver namespace/openshift-oauth-apiserver: Readiness probe status=failure output="Get \"https://10.217.0.28:8443/readyz\": context deadline exceeded" start-of-body= Dec 27 09:08:06 crc kubenswrapper[4934]: I1227 09:08:06.924495 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hzmws" podUID="145ee02a-c7aa-4990-a288-e6f69f8ab030" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.28:8443/readyz\": context deadline exceeded" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.001276 4934 patch_prober.go:28] interesting pod/perses-operator-5bf474d74f-tx6z5 container/perses-operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.13:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.001340 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" podUID="f97e8d86-1e9e-4c5a-a164-2600193c576b" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.13:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.042488 4934 patch_prober.go:28] interesting pod/console-74cc7c486-2dpkp container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.138:8443/health\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.042569 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-74cc7c486-2dpkp" podUID="e0978ad2-1582-49d4-b239-ff929dafb117" containerName="console" probeResult="failure" output="Get \"https://10.217.0.138:8443/health\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.083447 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" podUID="fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.93:8080/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.083566 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.143401 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc" podUID="b461380d-6573-4898-9bc8-c82ceba822d8" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.101:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.280716 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.427388 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" podUID="df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.94:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.427679 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.427558 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" podUID="df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.94:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.428028 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.429288 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="webhook-server" containerStatusID={"Type":"cri-o","ID":"0544eea826fad2f82fea05a46a3bf73c02516d04af2f770e270401a3e238142f"} pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" containerMessage="Container webhook-server failed liveness probe, will be restarted" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.429347 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" podUID="df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb" containerName="webhook-server" containerID="cri-o://0544eea826fad2f82fea05a46a3bf73c02516d04af2f770e270401a3e238142f" gracePeriod=2 Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.452287 4934 patch_prober.go:28] interesting pod/thanos-querier-8467887986-svb7h container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.71:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.452566 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-8467887986-svb7h" podUID="a07fdfc4-584a-4a39-b87f-ee78f9cac687" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.71:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.559754 4934 patch_prober.go:28] interesting pod/logging-loki-gateway-74dbf44c45-qd69g container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.559813 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" podUID="47cbee2e-c325-4a8d-a26d-94bced60ee78" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.53:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.559818 4934 patch_prober.go:28] interesting pod/logging-loki-gateway-74dbf44c45-qd69g container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.559886 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-74dbf44c45-qd69g" podUID="47cbee2e-c325-4a8d-a26d-94bced60ee78" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.53:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.600175 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.644025 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-engine-55c9c9fc75-s5tvt" podUID="bb1fd5db-ac54-4e34-9cb3-4caae5f67da8" containerName="heat-engine" probeResult="failure" output="command timed out" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.644035 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/heat-engine-55c9c9fc75-s5tvt" podUID="bb1fd5db-ac54-4e34-9cb3-4caae5f67da8" containerName="heat-engine" probeResult="failure" output="command timed out" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.874430 4934 patch_prober.go:28] interesting pod/downloads-7954f5f757-8msj4 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.874779 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-8msj4" podUID="a0d34185-bf08-456c-a5fc-3f0852d35a91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.874822 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-8msj4" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.875199 4934 patch_prober.go:28] interesting pod/downloads-7954f5f757-8msj4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.875224 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8msj4" podUID="a0d34185-bf08-456c-a5fc-3f0852d35a91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.875272 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-8msj4" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.883433 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"bbb8c8c732b24041df4c223b3b91a7741cd1534a879275067b7b9c3d51a9deb7"} pod="openshift-console/downloads-7954f5f757-8msj4" containerMessage="Container download-server failed liveness probe, will be restarted" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.883489 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-8msj4" podUID="a0d34185-bf08-456c-a5fc-3f0852d35a91" containerName="download-server" containerID="cri-o://bbb8c8c732b24041df4c223b3b91a7741cd1534a879275067b7b9c3d51a9deb7" gracePeriod=2 Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.933449 4934 patch_prober.go:28] interesting pod/logging-loki-gateway-74dbf44c45-4ll7v container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.933508 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" podUID="eac5b26c-d598-4c68-95b5-e2583b456af9" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.54:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.933635 4934 patch_prober.go:28] interesting pod/logging-loki-gateway-74dbf44c45-4ll7v container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:07 crc kubenswrapper[4934]: I1227 09:08:07.933674 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-74dbf44c45-4ll7v" podUID="eac5b26c-d598-4c68-95b5-e2583b456af9" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.54:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.018906 4934 generic.go:334] "Generic (PLEG): container finished" podID="a4612dd4-0e0d-4c38-8da5-8bc30dee7c12" containerID="67e9d2fa2404b9f860a2bbc81d950b59caee85958a0d2750c45a0fcbd374a629" exitCode=0 Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.018978 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" event={"ID":"a4612dd4-0e0d-4c38-8da5-8bc30dee7c12","Type":"ContainerDied","Data":"67e9d2fa2404b9f860a2bbc81d950b59caee85958a0d2750c45a0fcbd374a629"} Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.021192 4934 generic.go:334] "Generic (PLEG): container finished" podID="f2d5e345-1bd9-4313-bdae-7637cb7ed944" containerID="4d1d4d6175948e5f26e1a1bc49d40f0216381680eaf82c8d4d5a44a2bc8b57a3" exitCode=1 Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.021747 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj" event={"ID":"f2d5e345-1bd9-4313-bdae-7637cb7ed944","Type":"ContainerDied","Data":"4d1d4d6175948e5f26e1a1bc49d40f0216381680eaf82c8d4d5a44a2bc8b57a3"} Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.023282 4934 generic.go:334] "Generic (PLEG): container finished" podID="4e39152f-3592-4907-9b06-d13074aeb3c4" containerID="2437eadde08e055b94e28a357e0995921db59807bc010af89911716c3aabd6e0" exitCode=2 Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.023356 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4e39152f-3592-4907-9b06-d13074aeb3c4","Type":"ContainerDied","Data":"2437eadde08e055b94e28a357e0995921db59807bc010af89911716c3aabd6e0"} Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.025422 4934 scope.go:117] "RemoveContainer" containerID="4d1d4d6175948e5f26e1a1bc49d40f0216381680eaf82c8d4d5a44a2bc8b57a3" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.025826 4934 generic.go:334] "Generic (PLEG): container finished" podID="f1cce183-e0a8-4cf1-854e-7a60986f194f" containerID="6282fc1c681be3cc65189166b54cc594fee49777c42b330d5aae499cb8508585" exitCode=1 Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.025904 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" event={"ID":"f1cce183-e0a8-4cf1-854e-7a60986f194f","Type":"ContainerDied","Data":"6282fc1c681be3cc65189166b54cc594fee49777c42b330d5aae499cb8508585"} Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.026932 4934 scope.go:117] "RemoveContainer" containerID="6282fc1c681be3cc65189166b54cc594fee49777c42b330d5aae499cb8508585" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.030986 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/2.log" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.033648 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.035797 4934 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="6d66e1134058289fe715318d0983996f62e222c539a7a940eb1ab6feb1889204" exitCode=1 Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.035847 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"6d66e1134058289fe715318d0983996f62e222c539a7a940eb1ab6feb1889204"} Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.035882 4934 scope.go:117] "RemoveContainer" containerID="5fa4c8b6a0bef1daea360a8dfb6c71a018b73f127f8fc5c58ef3220b1d5335e7" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.036946 4934 scope.go:117] "RemoveContainer" containerID="6d66e1134058289fe715318d0983996f62e222c539a7a940eb1ab6feb1889204" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.091275 4934 patch_prober.go:28] interesting pod/console-operator-58897d9998-jckgt container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.091295 4934 patch_prober.go:28] interesting pod/console-operator-58897d9998-jckgt container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.091336 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-jckgt" podUID="753c599b-a401-481e-a5b3-e2dd6ed2c4e7" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.091351 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-jckgt" podUID="753c599b-a401-481e-a5b3-e2dd6ed2c4e7" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.091414 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-jckgt" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.091439 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console-operator/console-operator-58897d9998-jckgt" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.092404 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="console-operator" containerStatusID={"Type":"cri-o","ID":"266ca8fe5b2103126717fbf60060271bbdf278ba0fdb38735bae74d5e0d42317"} pod="openshift-console-operator/console-operator-58897d9998-jckgt" containerMessage="Container console-operator failed liveness probe, will be restarted" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.092447 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console-operator/console-operator-58897d9998-jckgt" podUID="753c599b-a401-481e-a5b3-e2dd6ed2c4e7" containerName="console-operator" containerID="cri-o://266ca8fe5b2103126717fbf60060271bbdf278ba0fdb38735bae74d5e0d42317" gracePeriod=30 Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.132468 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" podUID="fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.93:8080/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.239899 4934 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5fhq7 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.241197 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" podUID="4c71cabc-c13f-4ae8-a704-10efee323e35" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.306889 4934 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-5cc9p container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.306934 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" podUID="af023fee-eae1-427a-8d70-487588422316" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.306975 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.307100 4934 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-5cc9p container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.307125 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" podUID="af023fee-eae1-427a-8d70-487588422316" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.307166 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.308013 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="catalog-operator" containerStatusID={"Type":"cri-o","ID":"f60d8867faf70d855bf4d3de99d3499deb657ee9ea0285565ed1ff36290fb267"} pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" containerMessage="Container catalog-operator failed liveness probe, will be restarted" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.308052 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" podUID="af023fee-eae1-427a-8d70-487588422316" containerName="catalog-operator" containerID="cri-o://f60d8867faf70d855bf4d3de99d3499deb657ee9ea0285565ed1ff36290fb267" gracePeriod=30 Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.524720 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/controller-5bddd4b946-7c5ls" podUID="a8825499-558f-4b31-b09b-0dcc92662f65" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.96:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.524795 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/controller-5bddd4b946-7c5ls" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.524802 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" podUID="df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.94:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.525024 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/controller-5bddd4b946-7c5ls" podUID="a8825499-558f-4b31-b09b-0dcc92662f65" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.96:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.525099 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5bddd4b946-7c5ls" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.526817 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="controller" containerStatusID={"Type":"cri-o","ID":"e5c694ae386e0a214260677899c88bc6eaa638ec88f3199e925722a091724206"} pod="metallb-system/controller-5bddd4b946-7c5ls" containerMessage="Container controller failed liveness probe, will be restarted" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.526906 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/controller-5bddd4b946-7c5ls" podUID="a8825499-558f-4b31-b09b-0dcc92662f65" containerName="controller" containerID="cri-o://e5c694ae386e0a214260677899c88bc6eaa638ec88f3199e925722a091724206" gracePeriod=2 Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.647042 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-hrs9g" podUID="73bc5c15-5c9d-49ae-bcd9-853df82fabbf" containerName="registry-server" probeResult="failure" output="command timed out" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.673840 4934 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-r6jld container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.673889 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" podUID="4ea539b6-ffe0-4393-af8d-62579266697c" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.673965 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.673835 4934 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-r6jld container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.675712 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" podUID="4ea539b6-ffe0-4393-af8d-62579266697c" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.675778 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.688506 4934 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-t6q48 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.688578 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" podUID="873d8371-2f45-409a-b18d-f4c7dbdabeea" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.688644 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.690073 4934 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-t6q48 container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.690173 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" podUID="873d8371-2f45-409a-b18d-f4c7dbdabeea" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.690223 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" Dec 27 09:08:08 crc kubenswrapper[4934]: E1227 09:08:08.772524 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45f4ca46_4041_4960_8c05_a20bd2c90b90.slice/crio-5f66e5c0191c6ce691d9d5526a2dabacd95623f4549ca175a37bdd114a0a0a6d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45f4ca46_4041_4960_8c05_a20bd2c90b90.slice/crio-conmon-5f66e5c0191c6ce691d9d5526a2dabacd95623f4549ca175a37bdd114a0a0a6d.scope\": RecentStats: unable to find data in memory cache]" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.796221 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.914306 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-v2lll" podUID="c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.914403 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/frr-k8s-v2lll" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.915639 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="frr" containerStatusID={"Type":"cri-o","ID":"a23c34a111bcac9a8c777af6384e620fb75b23d9f584e874f91638de1d4a1977"} pod="metallb-system/frr-k8s-v2lll" containerMessage="Container frr failed liveness probe, will be restarted" Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.915739 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-v2lll" podUID="c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd" containerName="frr" containerID="cri-o://a23c34a111bcac9a8c777af6384e620fb75b23d9f584e874f91638de1d4a1977" gracePeriod=2 Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.915819 4934 patch_prober.go:28] interesting pod/console-operator-58897d9998-jckgt container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": EOF" start-of-body= Dec 27 09:08:08 crc kubenswrapper[4934]: I1227 09:08:08.915878 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-jckgt" podUID="753c599b-a401-481e-a5b3-e2dd6ed2c4e7" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": EOF" Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.047292 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/2.log" Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.050854 4934 generic.go:334] "Generic (PLEG): container finished" podID="e2db190a-b455-44fc-a43e-2677e5af27b2" containerID="c9c80f7f97536e24c1f853c65d024959296423cac6628aefcc66960fe180b7d6" exitCode=1 Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.050894 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" event={"ID":"e2db190a-b455-44fc-a43e-2677e5af27b2","Type":"ContainerDied","Data":"c9c80f7f97536e24c1f853c65d024959296423cac6628aefcc66960fe180b7d6"} Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.052531 4934 scope.go:117] "RemoveContainer" containerID="c9c80f7f97536e24c1f853c65d024959296423cac6628aefcc66960fe180b7d6" Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.054745 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-58897d9998-jckgt_753c599b-a401-481e-a5b3-e2dd6ed2c4e7/console-operator/0.log" Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.054790 4934 generic.go:334] "Generic (PLEG): container finished" podID="753c599b-a401-481e-a5b3-e2dd6ed2c4e7" containerID="266ca8fe5b2103126717fbf60060271bbdf278ba0fdb38735bae74d5e0d42317" exitCode=1 Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.054842 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-jckgt" event={"ID":"753c599b-a401-481e-a5b3-e2dd6ed2c4e7","Type":"ContainerDied","Data":"266ca8fe5b2103126717fbf60060271bbdf278ba0fdb38735bae74d5e0d42317"} Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.056699 4934 generic.go:334] "Generic (PLEG): container finished" podID="2a688ece-0e00-4241-b7dd-9dde25f0eaf1" containerID="fd15bc9e9543c8afdb985cb0426e5686fff142c76fd4188ddea84d8885e949cc" exitCode=0 Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.056757 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" event={"ID":"2a688ece-0e00-4241-b7dd-9dde25f0eaf1","Type":"ContainerDied","Data":"fd15bc9e9543c8afdb985cb0426e5686fff142c76fd4188ddea84d8885e949cc"} Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.058679 4934 generic.go:334] "Generic (PLEG): container finished" podID="d090706d-34a7-4c72-9cb9-ad0601db35a6" containerID="c498ab4a25c84f59bd3cc280f09132eeb7322b5390dd0cc7d9e2839317406180" exitCode=1 Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.058750 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" event={"ID":"d090706d-34a7-4c72-9cb9-ad0601db35a6","Type":"ContainerDied","Data":"c498ab4a25c84f59bd3cc280f09132eeb7322b5390dd0cc7d9e2839317406180"} Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.059727 4934 scope.go:117] "RemoveContainer" containerID="c498ab4a25c84f59bd3cc280f09132eeb7322b5390dd0cc7d9e2839317406180" Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.061498 4934 generic.go:334] "Generic (PLEG): container finished" podID="45f4ca46-4041-4960-8c05-a20bd2c90b90" containerID="5f66e5c0191c6ce691d9d5526a2dabacd95623f4549ca175a37bdd114a0a0a6d" exitCode=0 Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.061565 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" event={"ID":"45f4ca46-4041-4960-8c05-a20bd2c90b90","Type":"ContainerDied","Data":"5f66e5c0191c6ce691d9d5526a2dabacd95623f4549ca175a37bdd114a0a0a6d"} Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.064370 4934 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="80370fd8607ddb05cccbe7383b94c5a4c1a874299bbbc7f665805ae5edf73ef9" exitCode=0 Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.064439 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"80370fd8607ddb05cccbe7383b94c5a4c1a874299bbbc7f665805ae5edf73ef9"} Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.065493 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="packageserver" containerStatusID={"Type":"cri-o","ID":"4bf1d93325579a031c6fd1a9de6d9d6c91a118fcca02cf8d81a3e755107b73c5"} pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" containerMessage="Container packageserver failed liveness probe, will be restarted" Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.065563 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" podUID="873d8371-2f45-409a-b18d-f4c7dbdabeea" containerName="packageserver" containerID="cri-o://4bf1d93325579a031c6fd1a9de6d9d6c91a118fcca02cf8d81a3e755107b73c5" gracePeriod=30 Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.066476 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="olm-operator" containerStatusID={"Type":"cri-o","ID":"d2beced8cabeaaa6b7e2703afeac985c57d9358880695b3f97dfd7e847227b58"} pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" containerMessage="Container olm-operator failed liveness probe, will be restarted" Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.066521 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" podUID="4ea539b6-ffe0-4393-af8d-62579266697c" containerName="olm-operator" containerID="cri-o://d2beced8cabeaaa6b7e2703afeac985c57d9358880695b3f97dfd7e847227b58" gracePeriod=30 Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.214039 4934 patch_prober.go:28] interesting pod/oauth-openshift-fd9565bb5-mkr88 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.72:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.214355 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" podUID="772a3e80-411a-4c15-9e1c-b702392e13af" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.72:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.214440 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.214064 4934 patch_prober.go:28] interesting pod/oauth-openshift-fd9565bb5-mkr88 container/oauth-openshift namespace/openshift-authentication: Liveness probe status=failure output="Get \"https://10.217.0.72:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.214715 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" podUID="772a3e80-411a-4c15-9e1c-b702392e13af" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.72:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.214769 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.215390 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="oauth-openshift" containerStatusID={"Type":"cri-o","ID":"f9908286ac4156a77e85f0cdcafc047330f167accce9c72fcb833f0c42c2b04e"} pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" containerMessage="Container oauth-openshift failed liveness probe, will be restarted" Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.489957 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.693012 4934 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-t6q48 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.698407 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" podUID="873d8371-2f45-409a-b18d-f4c7dbdabeea" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.699364 4934 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-t6q48 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" start-of-body= Dec 27 09:08:09 crc kubenswrapper[4934]: I1227 09:08:09.699448 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" podUID="873d8371-2f45-409a-b18d-f4c7dbdabeea" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.114679 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" event={"ID":"45f4ca46-4041-4960-8c05-a20bd2c90b90","Type":"ContainerStarted","Data":"7156701865bcbe4e45f7419b4cbe625e8012f11f851ee20dfa2dae93a9bd01d9"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.117065 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.121938 4934 patch_prober.go:28] interesting pod/route-controller-manager-6bd9bdbfbf-x5vvz container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.77:8443/healthz\": dial tcp 10.217.0.77:8443: connect: connection refused" start-of-body= Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.121994 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" podUID="45f4ca46-4041-4960-8c05-a20bd2c90b90" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.77:8443/healthz\": dial tcp 10.217.0.77:8443: connect: connection refused" Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.122210 4934 generic.go:334] "Generic (PLEG): container finished" podID="964b7818-b04e-4c6d-8ad0-5a108513be73" containerID="bf4828a80aa2309175c3712a6b88e7388112737fa373bc6e393056a56a3ad929" exitCode=0 Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.122320 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" event={"ID":"964b7818-b04e-4c6d-8ad0-5a108513be73","Type":"ContainerDied","Data":"bf4828a80aa2309175c3712a6b88e7388112737fa373bc6e393056a56a3ad929"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.130119 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" event={"ID":"a4612dd4-0e0d-4c38-8da5-8bc30dee7c12","Type":"ContainerStarted","Data":"61642b10aafda10942d9561851424c13c8244a6a2402098bf7202038d485814d"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.130417 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.130824 4934 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5jzhr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.63:8080/healthz\": dial tcp 10.217.0.63:8080: connect: connection refused" start-of-body= Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.130857 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" podUID="a4612dd4-0e0d-4c38-8da5-8bc30dee7c12" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.63:8080/healthz\": dial tcp 10.217.0.63:8080: connect: connection refused" Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.134049 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.141340 4934 generic.go:334] "Generic (PLEG): container finished" podID="77029090-da44-403b-a20f-d72105281956" containerID="f2c7f56b2f98c403b81d7d2c5df60467754fd6024eb74338cb3630cb95e8ca4b" exitCode=1 Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.141418 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" event={"ID":"77029090-da44-403b-a20f-d72105281956","Type":"ContainerDied","Data":"f2c7f56b2f98c403b81d7d2c5df60467754fd6024eb74338cb3630cb95e8ca4b"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.142443 4934 scope.go:117] "RemoveContainer" containerID="f2c7f56b2f98c403b81d7d2c5df60467754fd6024eb74338cb3630cb95e8ca4b" Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.146700 4934 generic.go:334] "Generic (PLEG): container finished" podID="4ea539b6-ffe0-4393-af8d-62579266697c" containerID="d2beced8cabeaaa6b7e2703afeac985c57d9358880695b3f97dfd7e847227b58" exitCode=0 Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.146761 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" event={"ID":"4ea539b6-ffe0-4393-af8d-62579266697c","Type":"ContainerDied","Data":"d2beced8cabeaaa6b7e2703afeac985c57d9358880695b3f97dfd7e847227b58"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.151523 4934 generic.go:334] "Generic (PLEG): container finished" podID="1c0960f0-ef55-479a-b375-a09e69254743" containerID="9d102a6913460ad85934ce0bcd90edecc2fcb98d6472e06ef18cbbece71cf1bf" exitCode=1 Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.151632 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" event={"ID":"1c0960f0-ef55-479a-b375-a09e69254743","Type":"ContainerDied","Data":"9d102a6913460ad85934ce0bcd90edecc2fcb98d6472e06ef18cbbece71cf1bf"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.152398 4934 scope.go:117] "RemoveContainer" containerID="9d102a6913460ad85934ce0bcd90edecc2fcb98d6472e06ef18cbbece71cf1bf" Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.154038 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" event={"ID":"2a688ece-0e00-4241-b7dd-9dde25f0eaf1","Type":"ContainerStarted","Data":"ad49b5200e7a2d75b83e6ee25c7f5bfb4ec106429c3a54fad411aa93ddb5ee64"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.154881 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.155448 4934 patch_prober.go:28] interesting pod/controller-manager-7967777f79-kzj4d container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.83:8443/healthz\": dial tcp 10.217.0.83:8443: connect: connection refused" start-of-body= Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.155504 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" podUID="2a688ece-0e00-4241-b7dd-9dde25f0eaf1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.83:8443/healthz\": dial tcp 10.217.0.83:8443: connect: connection refused" Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.161281 4934 generic.go:334] "Generic (PLEG): container finished" podID="37d56061-eab6-43e5-8dec-f59220da5d0d" containerID="50262c84ec47924a26f26f43ac5a26fa2fc012c9e82081b4606dec8df1413648" exitCode=1 Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.161405 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cc599445b-dlwjl" event={"ID":"37d56061-eab6-43e5-8dec-f59220da5d0d","Type":"ContainerDied","Data":"50262c84ec47924a26f26f43ac5a26fa2fc012c9e82081b4606dec8df1413648"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.163666 4934 scope.go:117] "RemoveContainer" containerID="50262c84ec47924a26f26f43ac5a26fa2fc012c9e82081b4606dec8df1413648" Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.179559 4934 generic.go:334] "Generic (PLEG): container finished" podID="c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd" containerID="a23c34a111bcac9a8c777af6384e620fb75b23d9f584e874f91638de1d4a1977" exitCode=143 Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.179644 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2lll" event={"ID":"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd","Type":"ContainerDied","Data":"a23c34a111bcac9a8c777af6384e620fb75b23d9f584e874f91638de1d4a1977"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.204680 4934 generic.go:334] "Generic (PLEG): container finished" podID="35d6eb8d-3797-432c-a5ea-65f31adf7262" containerID="c637fdad1fae7799a7838d88477ef07194ee6d91bb834611b714d8bd903dab34" exitCode=1 Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.204810 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" event={"ID":"35d6eb8d-3797-432c-a5ea-65f31adf7262","Type":"ContainerDied","Data":"c637fdad1fae7799a7838d88477ef07194ee6d91bb834611b714d8bd903dab34"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.205766 4934 scope.go:117] "RemoveContainer" containerID="c637fdad1fae7799a7838d88477ef07194ee6d91bb834611b714d8bd903dab34" Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.215098 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" event={"ID":"f1cce183-e0a8-4cf1-854e-7a60986f194f","Type":"ContainerStarted","Data":"a28d99d64892fb62a635df812a027e680e82229255884f44170e307ca2b9cec4"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.215424 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.218510 4934 generic.go:334] "Generic (PLEG): container finished" podID="af023fee-eae1-427a-8d70-487588422316" containerID="f60d8867faf70d855bf4d3de99d3499deb657ee9ea0285565ed1ff36290fb267" exitCode=0 Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.218584 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" event={"ID":"af023fee-eae1-427a-8d70-487588422316","Type":"ContainerDied","Data":"f60d8867faf70d855bf4d3de99d3499deb657ee9ea0285565ed1ff36290fb267"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.220917 4934 generic.go:334] "Generic (PLEG): container finished" podID="b5ea792c-b43f-4221-b7bb-aba3421de3d8" containerID="75dfa517cd5c84973d7563d3a3f58980a7e1672fd90a8b7a567fefc59a74e1f8" exitCode=1 Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.220974 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" event={"ID":"b5ea792c-b43f-4221-b7bb-aba3421de3d8","Type":"ContainerDied","Data":"75dfa517cd5c84973d7563d3a3f58980a7e1672fd90a8b7a567fefc59a74e1f8"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.222020 4934 scope.go:117] "RemoveContainer" containerID="75dfa517cd5c84973d7563d3a3f58980a7e1672fd90a8b7a567fefc59a74e1f8" Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.226095 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj" event={"ID":"f2d5e345-1bd9-4313-bdae-7637cb7ed944","Type":"ContainerStarted","Data":"cb7fd5e07a318164e632bf76796d0b522a7c54ef1c41a1750279b3197dfbe293"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.226411 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj" Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.229191 4934 generic.go:334] "Generic (PLEG): container finished" podID="873d8371-2f45-409a-b18d-f4c7dbdabeea" containerID="4bf1d93325579a031c6fd1a9de6d9d6c91a118fcca02cf8d81a3e755107b73c5" exitCode=0 Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.229253 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" event={"ID":"873d8371-2f45-409a-b18d-f4c7dbdabeea","Type":"ContainerDied","Data":"4bf1d93325579a031c6fd1a9de6d9d6c91a118fcca02cf8d81a3e755107b73c5"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.233216 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8825499-558f-4b31-b09b-0dcc92662f65" containerID="e5c694ae386e0a214260677899c88bc6eaa638ec88f3199e925722a091724206" exitCode=0 Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.233276 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-7c5ls" event={"ID":"a8825499-558f-4b31-b09b-0dcc92662f65","Type":"ContainerDied","Data":"e5c694ae386e0a214260677899c88bc6eaa638ec88f3199e925722a091724206"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.236091 4934 generic.go:334] "Generic (PLEG): container finished" podID="df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb" containerID="0544eea826fad2f82fea05a46a3bf73c02516d04af2f770e270401a3e238142f" exitCode=0 Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.236136 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" event={"ID":"df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb","Type":"ContainerDied","Data":"0544eea826fad2f82fea05a46a3bf73c02516d04af2f770e270401a3e238142f"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.240899 4934 generic.go:334] "Generic (PLEG): container finished" podID="718122b6-c9f4-4d6f-92b5-6862c15f4ee6" containerID="ae6840f3f2f83c27940fbd80c73057b10eb4929c250ae9c6a06e465bba91b82c" exitCode=1 Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.240940 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" event={"ID":"718122b6-c9f4-4d6f-92b5-6862c15f4ee6","Type":"ContainerDied","Data":"ae6840f3f2f83c27940fbd80c73057b10eb4929c250ae9c6a06e465bba91b82c"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.241766 4934 scope.go:117] "RemoveContainer" containerID="ae6840f3f2f83c27940fbd80c73057b10eb4929c250ae9c6a06e465bba91b82c" Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.246189 4934 generic.go:334] "Generic (PLEG): container finished" podID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerID="9995c35f73f77d890191ce4b613fd43a168910498934a2d034fa50716244fb04" exitCode=0 Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.246245 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12e88dda-31ca-4b5b-93c3-7eced0cacda7","Type":"ContainerDied","Data":"9995c35f73f77d890191ce4b613fd43a168910498934a2d034fa50716244fb04"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.248813 4934 generic.go:334] "Generic (PLEG): container finished" podID="f41746ba-5695-41bc-8078-29ba3d80e3f1" containerID="ff80d6933782e700e7230918f89286a2cea0af4dd10686c7b85ecbd92b9f6cb3" exitCode=1 Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.248978 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" event={"ID":"f41746ba-5695-41bc-8078-29ba3d80e3f1","Type":"ContainerDied","Data":"ff80d6933782e700e7230918f89286a2cea0af4dd10686c7b85ecbd92b9f6cb3"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.250375 4934 scope.go:117] "RemoveContainer" containerID="ff80d6933782e700e7230918f89286a2cea0af4dd10686c7b85ecbd92b9f6cb3" Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.256648 4934 patch_prober.go:28] interesting pod/oauth-openshift-fd9565bb5-mkr88 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.72:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.256699 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" podUID="772a3e80-411a-4c15-9e1c-b702392e13af" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.72:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.261703 4934 generic.go:334] "Generic (PLEG): container finished" podID="4575d285-879f-4cc3-a9b9-406dc96e852d" containerID="0e9c0d7603353e305c0f51f83e3c3c23c80d0364f79fc2f47a4a05801de2061b" exitCode=0 Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.261775 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" event={"ID":"4575d285-879f-4cc3-a9b9-406dc96e852d","Type":"ContainerDied","Data":"0e9c0d7603353e305c0f51f83e3c3c23c80d0364f79fc2f47a4a05801de2061b"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.309222 4934 generic.go:334] "Generic (PLEG): container finished" podID="a0d34185-bf08-456c-a5fc-3f0852d35a91" containerID="bbb8c8c732b24041df4c223b3b91a7741cd1534a879275067b7b9c3d51a9deb7" exitCode=0 Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.309422 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8msj4" event={"ID":"a0d34185-bf08-456c-a5fc-3f0852d35a91","Type":"ContainerDied","Data":"bbb8c8c732b24041df4c223b3b91a7741cd1534a879275067b7b9c3d51a9deb7"} Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.409601 4934 patch_prober.go:28] interesting pod/loki-operator-controller-manager-77c7786d5-gz9d9 container/manager namespace/openshift-operators-redhat: Readiness probe status=failure output="Get \"http://10.217.0.47:8081/readyz\": dial tcp 10.217.0.47:8081: connect: connection refused" start-of-body= Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.409655 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" podUID="00799e47-9b94-49bf-8bd1-d6bb5036285a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.47:8081/readyz\": dial tcp 10.217.0.47:8081: connect: connection refused" Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.481230 4934 patch_prober.go:28] interesting pod/controller-manager-7967777f79-kzj4d container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.83:8443/healthz\": dial tcp 10.217.0.83:8443: connect: connection refused" start-of-body= Dec 27 09:08:10 crc kubenswrapper[4934]: I1227 09:08:10.481469 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" podUID="2a688ece-0e00-4241-b7dd-9dde25f0eaf1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.83:8443/healthz\": dial tcp 10.217.0.83:8443: connect: connection refused" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.081707 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-5f678c8dd6-nqqn7" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.244175 4934 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5fhq7 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.244458 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" podUID="4c71cabc-c13f-4ae8-a704-10efee323e35" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.301039 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-76788598db-6wrwh" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.302730 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-69d9546745-4b9rt" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.324577 4934 trace.go:236] Trace[439361213]: "Calculate volume metrics of storage for pod openshift-logging/logging-loki-index-gateway-0" (27-Dec-2025 09:08:09.407) (total time: 1917ms): Dec 27 09:08:11 crc kubenswrapper[4934]: Trace[439361213]: [1.917431785s] [1.917431785s] END Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.339665 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" event={"ID":"d090706d-34a7-4c72-9cb9-ad0601db35a6","Type":"ContainerStarted","Data":"8116d6f083bd3b1cf45a87b2392b4f3fbfd89f559e668058c87b8a9c2791b938"} Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.340793 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.344790 4934 trace.go:236] Trace[87766118]: "Calculate volume metrics of mysql-db for pod openstack/openstack-galera-0" (27-Dec-2025 09:08:08.801) (total time: 2543ms): Dec 27 09:08:11 crc kubenswrapper[4934]: Trace[87766118]: [2.543145983s] [2.543145983s] END Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.346857 4934 generic.go:334] "Generic (PLEG): container finished" podID="dc4ddef4-e9a5-4934-846e-1d1bcfe9f223" containerID="accb35096626791d61db24a47640ffc8de84625b32b0f51728a4076fcb71a160" exitCode=1 Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.347054 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" event={"ID":"dc4ddef4-e9a5-4934-846e-1d1bcfe9f223","Type":"ContainerDied","Data":"accb35096626791d61db24a47640ffc8de84625b32b0f51728a4076fcb71a160"} Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.349245 4934 scope.go:117] "RemoveContainer" containerID="accb35096626791d61db24a47640ffc8de84625b32b0f51728a4076fcb71a160" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.356556 4934 generic.go:334] "Generic (PLEG): container finished" podID="4c71cabc-c13f-4ae8-a704-10efee323e35" containerID="b631e608d82ae75f36a5a9fa3ea7f2bfcda667d3a9703769b16b2e70e44081be" exitCode=0 Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.356929 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" event={"ID":"4c71cabc-c13f-4ae8-a704-10efee323e35","Type":"ContainerDied","Data":"b631e608d82ae75f36a5a9fa3ea7f2bfcda667d3a9703769b16b2e70e44081be"} Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.360647 4934 generic.go:334] "Generic (PLEG): container finished" podID="59d65d09-c15c-49d9-8ee7-472ffe0dda55" containerID="fd3b03e8c2268a330af22ea6886e6379543ac0c2538b1bc390292173c7cb0d81" exitCode=1 Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.360694 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" event={"ID":"59d65d09-c15c-49d9-8ee7-472ffe0dda55","Type":"ContainerDied","Data":"fd3b03e8c2268a330af22ea6886e6379543ac0c2538b1bc390292173c7cb0d81"} Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.363716 4934 scope.go:117] "RemoveContainer" containerID="fd3b03e8c2268a330af22ea6886e6379543ac0c2538b1bc390292173c7cb0d81" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.374911 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/2.log" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.389070 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f1d3ac602edb9810f98f35dd2e1f3dc7b5107e9c91cb4b25e87ae92dce30f7b7"} Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.401368 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4e39152f-3592-4907-9b06-d13074aeb3c4","Type":"ContainerStarted","Data":"d0394d64dffdbcbc93f12b503d6bb7bc9633700b04c6c3331392b2d51a775a38"} Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.411873 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" event={"ID":"964b7818-b04e-4c6d-8ad0-5a108513be73","Type":"ContainerStarted","Data":"7d0304c25993607eb5e5eed524c4122307c74e0ee68c2c79e3cea97e713aba19"} Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.411920 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.412505 4934 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-2n6w2 container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" start-of-body= Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.412572 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" podUID="964b7818-b04e-4c6d-8ad0-5a108513be73" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.417536 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" event={"ID":"af023fee-eae1-427a-8d70-487588422316","Type":"ContainerStarted","Data":"e15cbc4b5c8ad4c920d8750fa92e9dd73909f84d43ced47b67e21c6c99f21975"} Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.422301 4934 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-5cc9p container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.422367 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" podUID="af023fee-eae1-427a-8d70-487588422316" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.433524 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7092457e9a1690496124671075c0b89db58d7fc4e0b125712140e8c0538a64e8"} Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.436957 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.444115 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8msj4" event={"ID":"a0d34185-bf08-456c-a5fc-3f0852d35a91","Type":"ContainerStarted","Data":"46f381ffe391f678de1c309a1ae90a444e3e40a01eb8b0fbd50260ceb127eca7"} Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.445752 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-8msj4" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.449315 4934 patch_prober.go:28] interesting pod/downloads-7954f5f757-8msj4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.449368 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8msj4" podUID="a0d34185-bf08-456c-a5fc-3f0852d35a91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.453827 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" event={"ID":"4ea539b6-ffe0-4393-af8d-62579266697c","Type":"ContainerStarted","Data":"9a8c07057032ae57e33975ceb6787e09fa0497c09fc28158e1ea5c3bde4f97cd"} Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.454066 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.473281 4934 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-r6jld container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" start-of-body= Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.473672 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" podUID="4ea539b6-ffe0-4393-af8d-62579266697c" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.519559 4934 generic.go:334] "Generic (PLEG): container finished" podID="9fbceb4d-e7ae-42c0-8a3f-17807bc9f2ca" containerID="0fb7ff640bccf350e32ade99fb8f1d4e2ae0f966f789bd42b1ff56f9ef695f59" exitCode=1 Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.526072 4934 patch_prober.go:28] interesting pod/controller-manager-7967777f79-kzj4d container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.83:8443/healthz\": dial tcp 10.217.0.83:8443: connect: connection refused" start-of-body= Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.526135 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" podUID="2a688ece-0e00-4241-b7dd-9dde25f0eaf1" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.83:8443/healthz\": dial tcp 10.217.0.83:8443: connect: connection refused" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.526329 4934 patch_prober.go:28] interesting pod/route-controller-manager-6bd9bdbfbf-x5vvz container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.77:8443/healthz\": dial tcp 10.217.0.77:8443: connect: connection refused" start-of-body= Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.526344 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" podUID="45f4ca46-4041-4960-8c05-a20bd2c90b90" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.77:8443/healthz\": dial tcp 10.217.0.77:8443: connect: connection refused" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.526339 4934 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5jzhr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.63:8080/healthz\": dial tcp 10.217.0.63:8080: connect: connection refused" start-of-body= Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.526371 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" podUID="a4612dd4-0e0d-4c38-8da5-8bc30dee7c12" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.63:8080/healthz\": dial tcp 10.217.0.63:8080: connect: connection refused" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.528022 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr" event={"ID":"9fbceb4d-e7ae-42c0-8a3f-17807bc9f2ca","Type":"ContainerDied","Data":"0fb7ff640bccf350e32ade99fb8f1d4e2ae0f966f789bd42b1ff56f9ef695f59"} Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.534348 4934 scope.go:117] "RemoveContainer" containerID="0fb7ff640bccf350e32ade99fb8f1d4e2ae0f966f789bd42b1ff56f9ef695f59" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.611669 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="ca95a4ef-282d-4c26-983f-d7d00f90443c" containerName="galera" containerID="cri-o://e862df5ec5f94dc038eceb2b1533dcf21f0a38c95f61e193874266056fd27165" gracePeriod=22 Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.639608 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="d03cf582-2192-4a10-b21e-10befebde419" containerName="galera" probeResult="failure" output="command timed out" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.639664 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/openstack-galera-0" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.640685 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="galera" containerStatusID={"Type":"cri-o","ID":"a04418c8ab5bb1fa91f94e0931bec31f25b20782638b4a478cac208f02a4620d"} pod="openstack/openstack-galera-0" containerMessage="Container galera failed liveness probe, will be restarted" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.641581 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="d03cf582-2192-4a10-b21e-10befebde419" containerName="galera" probeResult="failure" output="command timed out" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.641635 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.839937 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 27 09:08:11 crc kubenswrapper[4934]: I1227 09:08:11.869580 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-f8fb84555-fwjdt" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.085603 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.107244 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.140271 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.391612 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.391931 4934 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.391961 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.467661 4934 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5jzhr container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.63:8080/healthz\": dial tcp 10.217.0.63:8080: connect: connection refused" start-of-body= Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.467722 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" podUID="a4612dd4-0e0d-4c38-8da5-8bc30dee7c12" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.63:8080/healthz\": dial tcp 10.217.0.63:8080: connect: connection refused" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.468167 4934 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5jzhr container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.63:8080/healthz\": dial tcp 10.217.0.63:8080: connect: connection refused" start-of-body= Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.468204 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" podUID="a4612dd4-0e0d-4c38-8da5-8bc30dee7c12" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.63:8080/healthz\": dial tcp 10.217.0.63:8080: connect: connection refused" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.468767 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:08:12 crc kubenswrapper[4934]: E1227 09:08:12.469035 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.532560 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" event={"ID":"4575d285-879f-4cc3-a9b9-406dc96e852d","Type":"ContainerStarted","Data":"b0d34b57652be9649c746c6785b95f3a9fdcef7cab158c4baa336570fb19e13d"} Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.532871 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.532973 4934 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-lnknm container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.16:8081/healthz\": dial tcp 10.217.0.16:8081: connect: connection refused" start-of-body= Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.533017 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" podUID="4575d285-879f-4cc3-a9b9-406dc96e852d" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.16:8081/healthz\": dial tcp 10.217.0.16:8081: connect: connection refused" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.538606 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-5bddd4b946-7c5ls" event={"ID":"a8825499-558f-4b31-b09b-0dcc92662f65","Type":"ContainerStarted","Data":"57b7d63e5cdc272f0e3d35ebcb0278a3a06e14eafd26f95091179bc32f9110e3"} Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.538692 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-5bddd4b946-7c5ls" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.541998 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" event={"ID":"e2db190a-b455-44fc-a43e-2677e5af27b2","Type":"ContainerStarted","Data":"77e72f5d1c2d07726a11ee380f6984b3af9220eaa210d9319bec7ba48f820363"} Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.542456 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.547382 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" event={"ID":"b5ea792c-b43f-4221-b7bb-aba3421de3d8","Type":"ContainerStarted","Data":"f006cab213807f9d44d6152c6e0901c4cb50082484283547594c16de26e29450"} Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.547731 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.550040 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-58897d9998-jckgt_753c599b-a401-481e-a5b3-e2dd6ed2c4e7/console-operator/0.log" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.550111 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-jckgt" event={"ID":"753c599b-a401-481e-a5b3-e2dd6ed2c4e7","Type":"ContainerStarted","Data":"1b163a23a6c1ba86baea1876f0046f03aabab7c4cbad6f38b1b00a2518b14c66"} Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.551688 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-jckgt" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.551800 4934 patch_prober.go:28] interesting pod/console-operator-58897d9998-jckgt container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.551850 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-jckgt" podUID="753c599b-a401-481e-a5b3-e2dd6ed2c4e7" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.558431 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" event={"ID":"873d8371-2f45-409a-b18d-f4c7dbdabeea","Type":"ContainerStarted","Data":"8623675cefc70db33fb34773f131af39ccc6958d318e7943abaadc85685e5f78"} Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.560798 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.560874 4934 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-t6q48 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" start-of-body= Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.560900 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" podUID="873d8371-2f45-409a-b18d-f4c7dbdabeea" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.570447 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cc599445b-dlwjl" event={"ID":"37d56061-eab6-43e5-8dec-f59220da5d0d","Type":"ContainerStarted","Data":"a081989a691a1ace5d52a7039d8084d9b883a372c5cf532eab5b80d49097bbfc"} Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.571459 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7cc599445b-dlwjl" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.588011 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-v2lll" event={"ID":"c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd","Type":"ContainerStarted","Data":"5e5c4110b1e38af977c6dc2da8282b5adf191228f5410ee70bd90cc46ce1d71d"} Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.590071 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.590272 4934 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-2n6w2 container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" start-of-body= Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.590310 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" podUID="964b7818-b04e-4c6d-8ad0-5a108513be73" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.590615 4934 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-r6jld container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" start-of-body= Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.590662 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" podUID="4ea539b6-ffe0-4393-af8d-62579266697c" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.592210 4934 patch_prober.go:28] interesting pod/downloads-7954f5f757-8msj4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.592241 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8msj4" podUID="a0d34185-bf08-456c-a5fc-3f0852d35a91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.592308 4934 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-5cc9p container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.592897 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" podUID="af023fee-eae1-427a-8d70-487588422316" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.620253 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.702187 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-64fb555449-tbmpv" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.754351 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.782198 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6bd9bdbfbf-x5vvz" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.829934 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="d03cf582-2192-4a10-b21e-10befebde419" containerName="galera" containerID="cri-o://a04418c8ab5bb1fa91f94e0931bec31f25b20782638b4a478cac208f02a4620d" gracePeriod=29 Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.870891 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-v2lll" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.923825 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.924159 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr" Dec 27 09:08:12 crc kubenswrapper[4934]: I1227 09:08:12.949695 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-v2lll" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.089683 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.089733 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.092692 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.107753 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-64d7c556cd-ffx59" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.236917 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.311040 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.394532 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-7c9ff8845d-ghvpg" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.603530 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr" event={"ID":"9fbceb4d-e7ae-42c0-8a3f-17807bc9f2ca","Type":"ContainerStarted","Data":"3efb435fac76eca3521088f4ca40be0c177ed531cce2f3f36f31e0b61924c65a"} Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.603665 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.605770 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" event={"ID":"f41746ba-5695-41bc-8078-29ba3d80e3f1","Type":"ContainerStarted","Data":"6061579c67de399392fa5cfcdd76c49e61ef43e2582598a7596da7f979f656ad"} Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.606349 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.608982 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" event={"ID":"59d65d09-c15c-49d9-8ee7-472ffe0dda55","Type":"ContainerStarted","Data":"7ddeb28247a8f4e7545d37e0fa49f0602ec9a3f1feb7e060c409f3e5e55a1e19"} Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.609586 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.611887 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" event={"ID":"77029090-da44-403b-a20f-d72105281956","Type":"ContainerStarted","Data":"cd0b904534bfddff6d62d1c1e3131d7bf30e12ccceaf264193b9b97bf2e534dc"} Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.612524 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.615003 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" event={"ID":"dc4ddef4-e9a5-4934-846e-1d1bcfe9f223","Type":"ContainerStarted","Data":"0e2723a2878d45478b58443e2f9b08633fb726a927edbf393519ed7807f3ae5d"} Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.615159 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.617757 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" event={"ID":"4c71cabc-c13f-4ae8-a704-10efee323e35","Type":"ContainerStarted","Data":"3568fc3b6f0560eaa182601df71838bbcdfb0eca7527c2ec3a2fa30d99bbe793"} Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.617944 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.620014 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" event={"ID":"35d6eb8d-3797-432c-a5ea-65f31adf7262","Type":"ContainerStarted","Data":"3667482b0f52168edbc623ac316981d182217d6209416d32d86e5b06af0cc182"} Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.620413 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.622063 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" event={"ID":"718122b6-c9f4-4d6f-92b5-6862c15f4ee6","Type":"ContainerStarted","Data":"34e7d91d91739ccd3db54e7e84ea15a1ee90df3645855b7e4621e172a4aa4bbe"} Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.622246 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.624503 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" event={"ID":"df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb","Type":"ContainerStarted","Data":"ea7c2a4006fa18d9a4cf896be397c64275d7ad48f7297d93eb13f7dd7069b30e"} Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.624669 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.626870 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" event={"ID":"1c0960f0-ef55-479a-b375-a09e69254743","Type":"ContainerStarted","Data":"e8ad0f63586d7b61b4ac19b34910d1c56d6758b78b1302387c0a38aeb6658167"} Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.627311 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.630877 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12e88dda-31ca-4b5b-93c3-7eced0cacda7","Type":"ContainerStarted","Data":"4b06d76681d3d075d4efcabe254bf4613f7808f693f7a1abb914d26a80a561cd"} Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.631835 4934 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-lnknm container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.16:8081/healthz\": dial tcp 10.217.0.16:8081: connect: connection refused" start-of-body= Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.631860 4934 patch_prober.go:28] interesting pod/console-operator-58897d9998-jckgt container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.631869 4934 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-t6q48 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" start-of-body= Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.631876 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" podUID="4575d285-879f-4cc3-a9b9-406dc96e852d" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.16:8081/healthz\": dial tcp 10.217.0.16:8081: connect: connection refused" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.631893 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-jckgt" podUID="753c599b-a401-481e-a5b3-e2dd6ed2c4e7" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.631899 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" podUID="873d8371-2f45-409a-b18d-f4c7dbdabeea" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.631975 4934 patch_prober.go:28] interesting pod/downloads-7954f5f757-8msj4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.632008 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8msj4" podUID="a0d34185-bf08-456c-a5fc-3f0852d35a91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.632252 4934 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-5cc9p container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.632304 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" podUID="af023fee-eae1-427a-8d70-487588422316" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.784423 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-2n6w2" Dec 27 09:08:13 crc kubenswrapper[4934]: I1227 09:08:13.796228 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5758858f4d-mg9wf" Dec 27 09:08:14 crc kubenswrapper[4934]: I1227 09:08:14.208485 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hrs9g" Dec 27 09:08:14 crc kubenswrapper[4934]: I1227 09:08:14.258691 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hrs9g" Dec 27 09:08:14 crc kubenswrapper[4934]: I1227 09:08:14.642125 4934 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-t6q48 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" start-of-body= Dec 27 09:08:14 crc kubenswrapper[4934]: I1227 09:08:14.642175 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" podUID="873d8371-2f45-409a-b18d-f4c7dbdabeea" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.35:5443/healthz\": dial tcp 10.217.0.35:5443: connect: connection refused" Dec 27 09:08:14 crc kubenswrapper[4934]: I1227 09:08:14.669661 4934 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-lnknm container/operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.16:8081/healthz\": dial tcp 10.217.0.16:8081: connect: connection refused" start-of-body= Dec 27 09:08:14 crc kubenswrapper[4934]: I1227 09:08:14.669705 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" podUID="4575d285-879f-4cc3-a9b9-406dc96e852d" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.16:8081/healthz\": dial tcp 10.217.0.16:8081: connect: connection refused" Dec 27 09:08:14 crc kubenswrapper[4934]: I1227 09:08:14.672329 4934 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-lnknm container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.16:8081/healthz\": dial tcp 10.217.0.16:8081: connect: connection refused" start-of-body= Dec 27 09:08:14 crc kubenswrapper[4934]: I1227 09:08:14.672546 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" podUID="4575d285-879f-4cc3-a9b9-406dc96e852d" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.16:8081/healthz\": dial tcp 10.217.0.16:8081: connect: connection refused" Dec 27 09:08:14 crc kubenswrapper[4934]: I1227 09:08:14.837011 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-tx6z5" Dec 27 09:08:14 crc kubenswrapper[4934]: I1227 09:08:14.989249 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-74cc7c486-2dpkp" Dec 27 09:08:15 crc kubenswrapper[4934]: I1227 09:08:15.074455 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-78479c6bd7-j8dhc" Dec 27 09:08:16 crc kubenswrapper[4934]: I1227 09:08:16.027544 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="46b6dbac-6489-4651-8956-8c8a9767a4aa" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 09:08:16 crc kubenswrapper[4934]: I1227 09:08:16.074006 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5b75b656f5-sqn2z" Dec 27 09:08:16 crc kubenswrapper[4934]: I1227 09:08:16.353179 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 09:08:16 crc kubenswrapper[4934]: I1227 09:08:16.713027 4934 generic.go:334] "Generic (PLEG): container finished" podID="ca95a4ef-282d-4c26-983f-d7d00f90443c" containerID="e862df5ec5f94dc038eceb2b1533dcf21f0a38c95f61e193874266056fd27165" exitCode=0 Dec 27 09:08:16 crc kubenswrapper[4934]: I1227 09:08:16.713069 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ca95a4ef-282d-4c26-983f-d7d00f90443c","Type":"ContainerDied","Data":"e862df5ec5f94dc038eceb2b1533dcf21f0a38c95f61e193874266056fd27165"} Dec 27 09:08:16 crc kubenswrapper[4934]: I1227 09:08:16.791552 4934 patch_prober.go:28] interesting pod/downloads-7954f5f757-8msj4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 27 09:08:16 crc kubenswrapper[4934]: I1227 09:08:16.791609 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8msj4" podUID="a0d34185-bf08-456c-a5fc-3f0852d35a91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 27 09:08:16 crc kubenswrapper[4934]: I1227 09:08:16.791527 4934 patch_prober.go:28] interesting pod/downloads-7954f5f757-8msj4 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 27 09:08:16 crc kubenswrapper[4934]: I1227 09:08:16.791702 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-8msj4" podUID="a0d34185-bf08-456c-a5fc-3f0852d35a91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 27 09:08:17 crc kubenswrapper[4934]: I1227 09:08:17.080327 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-jckgt" Dec 27 09:08:17 crc kubenswrapper[4934]: I1227 09:08:17.240037 4934 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5fhq7 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Dec 27 09:08:17 crc kubenswrapper[4934]: I1227 09:08:17.240104 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" podUID="4c71cabc-c13f-4ae8-a704-10efee323e35" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Dec 27 09:08:17 crc kubenswrapper[4934]: I1227 09:08:17.240122 4934 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5fhq7 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Dec 27 09:08:17 crc kubenswrapper[4934]: I1227 09:08:17.240216 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" podUID="4c71cabc-c13f-4ae8-a704-10efee323e35" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Dec 27 09:08:17 crc kubenswrapper[4934]: I1227 09:08:17.308425 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-5cc9p" Dec 27 09:08:17 crc kubenswrapper[4934]: I1227 09:08:17.517143 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="46b6dbac-6489-4651-8956-8c8a9767a4aa" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 09:08:17 crc kubenswrapper[4934]: I1227 09:08:17.696618 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-t6q48" Dec 27 09:08:17 crc kubenswrapper[4934]: I1227 09:08:17.709720 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-r6jld" Dec 27 09:08:17 crc kubenswrapper[4934]: I1227 09:08:17.767270 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ca95a4ef-282d-4c26-983f-d7d00f90443c","Type":"ContainerStarted","Data":"1125d13d84845c8bc5a56864895d5118a46d7eb79a7ce66971ab1297d8eeea94"} Dec 27 09:08:18 crc kubenswrapper[4934]: I1227 09:08:18.220158 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 09:08:18 crc kubenswrapper[4934]: I1227 09:08:18.784226 4934 generic.go:334] "Generic (PLEG): container finished" podID="d03cf582-2192-4a10-b21e-10befebde419" containerID="a04418c8ab5bb1fa91f94e0931bec31f25b20782638b4a478cac208f02a4620d" exitCode=0 Dec 27 09:08:18 crc kubenswrapper[4934]: I1227 09:08:18.786148 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d03cf582-2192-4a10-b21e-10befebde419","Type":"ContainerDied","Data":"a04418c8ab5bb1fa91f94e0931bec31f25b20782638b4a478cac208f02a4620d"} Dec 27 09:08:18 crc kubenswrapper[4934]: I1227 09:08:18.786221 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d03cf582-2192-4a10-b21e-10befebde419","Type":"ContainerStarted","Data":"561e17e787d762b518f758a4fa5ab27e6f9163d79fff09ac8a1aac1ce7341bda"} Dec 27 09:08:19 crc kubenswrapper[4934]: I1227 09:08:19.867801 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hrs9g"] Dec 27 09:08:19 crc kubenswrapper[4934]: I1227 09:08:19.869531 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hrs9g" podUID="73bc5c15-5c9d-49ae-bcd9-853df82fabbf" containerName="registry-server" containerID="cri-o://23bb638c28ceb00176e086fce835249595930f3094fbf84d7920bc0e0e3d1de1" gracePeriod=2 Dec 27 09:08:20 crc kubenswrapper[4934]: I1227 09:08:20.031760 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 27 09:08:20 crc kubenswrapper[4934]: I1227 09:08:20.033297 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 27 09:08:20 crc kubenswrapper[4934]: I1227 09:08:20.245126 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5fhq7" Dec 27 09:08:20 crc kubenswrapper[4934]: I1227 09:08:20.403381 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-77c7786d5-gz9d9" Dec 27 09:08:20 crc kubenswrapper[4934]: I1227 09:08:20.499852 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7967777f79-kzj4d" Dec 27 09:08:20 crc kubenswrapper[4934]: I1227 09:08:20.537801 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="46b6dbac-6489-4651-8956-8c8a9767a4aa" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 09:08:20 crc kubenswrapper[4934]: I1227 09:08:20.537881 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 27 09:08:20 crc kubenswrapper[4934]: I1227 09:08:20.539016 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cinder-scheduler" containerStatusID={"Type":"cri-o","ID":"d56b14937b8f4e2616b25041792887ddd2d75019777c58f0718b35781615cd6b"} pod="openstack/cinder-scheduler-0" containerMessage="Container cinder-scheduler failed liveness probe, will be restarted" Dec 27 09:08:20 crc kubenswrapper[4934]: I1227 09:08:20.539072 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="46b6dbac-6489-4651-8956-8c8a9767a4aa" containerName="cinder-scheduler" containerID="cri-o://d56b14937b8f4e2616b25041792887ddd2d75019777c58f0718b35781615cd6b" gracePeriod=30 Dec 27 09:08:21 crc kubenswrapper[4934]: I1227 09:08:21.371302 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 27 09:08:21 crc kubenswrapper[4934]: I1227 09:08:21.372325 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 27 09:08:21 crc kubenswrapper[4934]: I1227 09:08:21.901546 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 27 09:08:22 crc kubenswrapper[4934]: I1227 09:08:22.391809 4934 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 27 09:08:22 crc kubenswrapper[4934]: I1227 09:08:22.391887 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 27 09:08:22 crc kubenswrapper[4934]: I1227 09:08:22.471254 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5jzhr" Dec 27 09:08:22 crc kubenswrapper[4934]: I1227 09:08:22.843356 4934 generic.go:334] "Generic (PLEG): container finished" podID="73bc5c15-5c9d-49ae-bcd9-853df82fabbf" containerID="23bb638c28ceb00176e086fce835249595930f3094fbf84d7920bc0e0e3d1de1" exitCode=0 Dec 27 09:08:22 crc kubenswrapper[4934]: I1227 09:08:22.843431 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrs9g" event={"ID":"73bc5c15-5c9d-49ae-bcd9-853df82fabbf","Type":"ContainerDied","Data":"23bb638c28ceb00176e086fce835249595930f3094fbf84d7920bc0e0e3d1de1"} Dec 27 09:08:23 crc kubenswrapper[4934]: I1227 09:08:23.483043 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-649cf68dc9-t4bqz" Dec 27 09:08:23 crc kubenswrapper[4934]: I1227 09:08:23.484359 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-5b67cfc8fb-6r7n4" Dec 27 09:08:23 crc kubenswrapper[4934]: I1227 09:08:23.484488 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7f764db9b-pgjkp" Dec 27 09:08:23 crc kubenswrapper[4934]: I1227 09:08:23.486359 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-65cdd6cc59-ggc7k" Dec 27 09:08:23 crc kubenswrapper[4934]: I1227 09:08:23.486410 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-54996ddd58-rjkpr" Dec 27 09:08:23 crc kubenswrapper[4934]: I1227 09:08:23.492556 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-7564bdd644-bg78w" Dec 27 09:08:23 crc kubenswrapper[4934]: I1227 09:08:23.492649 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-6c77d5f585-lg74h" Dec 27 09:08:23 crc kubenswrapper[4934]: I1227 09:08:23.492688 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-6b444986fd-djrtj" Dec 27 09:08:23 crc kubenswrapper[4934]: I1227 09:08:23.563916 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-6659cb75fc-9kzpj" Dec 27 09:08:23 crc kubenswrapper[4934]: I1227 09:08:23.589050 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7cc599445b-dlwjl" Dec 27 09:08:23 crc kubenswrapper[4934]: I1227 09:08:23.592964 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-d5fb87cb8-w2xgr" Dec 27 09:08:23 crc kubenswrapper[4934]: I1227 09:08:23.605978 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5d79c6465c-zldvm" Dec 27 09:08:23 crc kubenswrapper[4934]: I1227 09:08:23.704032 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-59cf64b5bd-28lln" Dec 27 09:08:23 crc kubenswrapper[4934]: E1227 09:08:23.730268 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 23bb638c28ceb00176e086fce835249595930f3094fbf84d7920bc0e0e3d1de1 is running failed: container process not found" containerID="23bb638c28ceb00176e086fce835249595930f3094fbf84d7920bc0e0e3d1de1" cmd=["grpc_health_probe","-addr=:50051"] Dec 27 09:08:23 crc kubenswrapper[4934]: E1227 09:08:23.730593 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 23bb638c28ceb00176e086fce835249595930f3094fbf84d7920bc0e0e3d1de1 is running failed: container process not found" containerID="23bb638c28ceb00176e086fce835249595930f3094fbf84d7920bc0e0e3d1de1" cmd=["grpc_health_probe","-addr=:50051"] Dec 27 09:08:23 crc kubenswrapper[4934]: E1227 09:08:23.730907 4934 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 23bb638c28ceb00176e086fce835249595930f3094fbf84d7920bc0e0e3d1de1 is running failed: container process not found" containerID="23bb638c28ceb00176e086fce835249595930f3094fbf84d7920bc0e0e3d1de1" cmd=["grpc_health_probe","-addr=:50051"] Dec 27 09:08:23 crc kubenswrapper[4934]: E1227 09:08:23.730940 4934 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 23bb638c28ceb00176e086fce835249595930f3094fbf84d7920bc0e0e3d1de1 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-hrs9g" podUID="73bc5c15-5c9d-49ae-bcd9-853df82fabbf" containerName="registry-server" Dec 27 09:08:23 crc kubenswrapper[4934]: I1227 09:08:23.952645 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc" Dec 27 09:08:24 crc kubenswrapper[4934]: I1227 09:08:24.676656 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-lnknm" Dec 27 09:08:24 crc kubenswrapper[4934]: I1227 09:08:24.873449 4934 generic.go:334] "Generic (PLEG): container finished" podID="46b6dbac-6489-4651-8956-8c8a9767a4aa" containerID="d56b14937b8f4e2616b25041792887ddd2d75019777c58f0718b35781615cd6b" exitCode=0 Dec 27 09:08:24 crc kubenswrapper[4934]: I1227 09:08:24.873706 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"46b6dbac-6489-4651-8956-8c8a9767a4aa","Type":"ContainerDied","Data":"d56b14937b8f4e2616b25041792887ddd2d75019777c58f0718b35781615cd6b"} Dec 27 09:08:25 crc kubenswrapper[4934]: I1227 09:08:25.330461 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hrs9g" Dec 27 09:08:25 crc kubenswrapper[4934]: I1227 09:08:25.449054 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6rkd\" (UniqueName: \"kubernetes.io/projected/73bc5c15-5c9d-49ae-bcd9-853df82fabbf-kube-api-access-j6rkd\") pod \"73bc5c15-5c9d-49ae-bcd9-853df82fabbf\" (UID: \"73bc5c15-5c9d-49ae-bcd9-853df82fabbf\") " Dec 27 09:08:25 crc kubenswrapper[4934]: I1227 09:08:25.449178 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73bc5c15-5c9d-49ae-bcd9-853df82fabbf-catalog-content\") pod \"73bc5c15-5c9d-49ae-bcd9-853df82fabbf\" (UID: \"73bc5c15-5c9d-49ae-bcd9-853df82fabbf\") " Dec 27 09:08:25 crc kubenswrapper[4934]: I1227 09:08:25.449325 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73bc5c15-5c9d-49ae-bcd9-853df82fabbf-utilities\") pod \"73bc5c15-5c9d-49ae-bcd9-853df82fabbf\" (UID: \"73bc5c15-5c9d-49ae-bcd9-853df82fabbf\") " Dec 27 09:08:25 crc kubenswrapper[4934]: I1227 09:08:25.450490 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73bc5c15-5c9d-49ae-bcd9-853df82fabbf-utilities" (OuterVolumeSpecName: "utilities") pod "73bc5c15-5c9d-49ae-bcd9-853df82fabbf" (UID: "73bc5c15-5c9d-49ae-bcd9-853df82fabbf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:08:25 crc kubenswrapper[4934]: I1227 09:08:25.470978 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:08:25 crc kubenswrapper[4934]: E1227 09:08:25.471513 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:08:25 crc kubenswrapper[4934]: I1227 09:08:25.477947 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73bc5c15-5c9d-49ae-bcd9-853df82fabbf-kube-api-access-j6rkd" (OuterVolumeSpecName: "kube-api-access-j6rkd") pod "73bc5c15-5c9d-49ae-bcd9-853df82fabbf" (UID: "73bc5c15-5c9d-49ae-bcd9-853df82fabbf"). InnerVolumeSpecName "kube-api-access-j6rkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:08:25 crc kubenswrapper[4934]: I1227 09:08:25.538957 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73bc5c15-5c9d-49ae-bcd9-853df82fabbf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73bc5c15-5c9d-49ae-bcd9-853df82fabbf" (UID: "73bc5c15-5c9d-49ae-bcd9-853df82fabbf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:08:25 crc kubenswrapper[4934]: I1227 09:08:25.552002 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73bc5c15-5c9d-49ae-bcd9-853df82fabbf-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 09:08:25 crc kubenswrapper[4934]: I1227 09:08:25.552307 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6rkd\" (UniqueName: \"kubernetes.io/projected/73bc5c15-5c9d-49ae-bcd9-853df82fabbf-kube-api-access-j6rkd\") on node \"crc\" DevicePath \"\"" Dec 27 09:08:25 crc kubenswrapper[4934]: I1227 09:08:25.552394 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73bc5c15-5c9d-49ae-bcd9-853df82fabbf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 09:08:25 crc kubenswrapper[4934]: I1227 09:08:25.891597 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrs9g" event={"ID":"73bc5c15-5c9d-49ae-bcd9-853df82fabbf","Type":"ContainerDied","Data":"e3a9ef9e90cf6f45c7368ba388808bfbff8b7672e6a124fbda7635c6f8dc1365"} Dec 27 09:08:25 crc kubenswrapper[4934]: I1227 09:08:25.891995 4934 scope.go:117] "RemoveContainer" containerID="23bb638c28ceb00176e086fce835249595930f3094fbf84d7920bc0e0e3d1de1" Dec 27 09:08:26 crc kubenswrapper[4934]: I1227 09:08:25.891639 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hrs9g" Dec 27 09:08:26 crc kubenswrapper[4934]: I1227 09:08:26.200413 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hrs9g"] Dec 27 09:08:26 crc kubenswrapper[4934]: I1227 09:08:26.207286 4934 scope.go:117] "RemoveContainer" containerID="ed04061c0c586c29911ef265fb4612200b6233b0f439ff8b3b2c07885d724f48" Dec 27 09:08:26 crc kubenswrapper[4934]: I1227 09:08:26.218165 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hrs9g"] Dec 27 09:08:26 crc kubenswrapper[4934]: I1227 09:08:26.236218 4934 scope.go:117] "RemoveContainer" containerID="e0f3661839b28d617baf5eea31c844302cea31ec2a1859bff75b55bd5dfa0bd3" Dec 27 09:08:26 crc kubenswrapper[4934]: I1227 09:08:26.393451 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-8b4577cdd-qcwh8" Dec 27 09:08:26 crc kubenswrapper[4934]: I1227 09:08:26.791782 4934 patch_prober.go:28] interesting pod/downloads-7954f5f757-8msj4 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 27 09:08:26 crc kubenswrapper[4934]: I1227 09:08:26.792340 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-8msj4" podUID="a0d34185-bf08-456c-a5fc-3f0852d35a91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 27 09:08:26 crc kubenswrapper[4934]: I1227 09:08:26.791838 4934 patch_prober.go:28] interesting pod/downloads-7954f5f757-8msj4 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Dec 27 09:08:26 crc kubenswrapper[4934]: I1227 09:08:26.792526 4934 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8msj4" podUID="a0d34185-bf08-456c-a5fc-3f0852d35a91" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Dec 27 09:08:26 crc kubenswrapper[4934]: I1227 09:08:26.922071 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"46b6dbac-6489-4651-8956-8c8a9767a4aa","Type":"ContainerStarted","Data":"d26d96442fc0580edf0bce22c9b74c370ea1abc1d5b85a4a11b87445dc80c962"} Dec 27 09:08:27 crc kubenswrapper[4934]: I1227 09:08:27.398811 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-5bddd4b946-7c5ls" Dec 27 09:08:27 crc kubenswrapper[4934]: I1227 09:08:27.480923 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73bc5c15-5c9d-49ae-bcd9-853df82fabbf" path="/var/lib/kubelet/pods/73bc5c15-5c9d-49ae-bcd9-853df82fabbf/volumes" Dec 27 09:08:31 crc kubenswrapper[4934]: I1227 09:08:31.490641 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 27 09:08:31 crc kubenswrapper[4934]: I1227 09:08:31.505162 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="46b6dbac-6489-4651-8956-8c8a9767a4aa" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 09:08:32 crc kubenswrapper[4934]: I1227 09:08:32.392109 4934 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 27 09:08:32 crc kubenswrapper[4934]: I1227 09:08:32.393349 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 27 09:08:32 crc kubenswrapper[4934]: I1227 09:08:32.393422 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 09:08:32 crc kubenswrapper[4934]: I1227 09:08:32.394504 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"f1d3ac602edb9810f98f35dd2e1f3dc7b5107e9c91cb4b25e87ae92dce30f7b7"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Dec 27 09:08:32 crc kubenswrapper[4934]: I1227 09:08:32.394649 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://f1d3ac602edb9810f98f35dd2e1f3dc7b5107e9c91cb4b25e87ae92dce30f7b7" gracePeriod=30 Dec 27 09:08:34 crc kubenswrapper[4934]: I1227 09:08:34.249684 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" podUID="772a3e80-411a-4c15-9e1c-b702392e13af" containerName="oauth-openshift" containerID="cri-o://f9908286ac4156a77e85f0cdcafc047330f167accce9c72fcb833f0c42c2b04e" gracePeriod=15 Dec 27 09:08:35 crc kubenswrapper[4934]: I1227 09:08:35.008701 4934 generic.go:334] "Generic (PLEG): container finished" podID="772a3e80-411a-4c15-9e1c-b702392e13af" containerID="f9908286ac4156a77e85f0cdcafc047330f167accce9c72fcb833f0c42c2b04e" exitCode=0 Dec 27 09:08:35 crc kubenswrapper[4934]: I1227 09:08:35.008815 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" event={"ID":"772a3e80-411a-4c15-9e1c-b702392e13af","Type":"ContainerDied","Data":"f9908286ac4156a77e85f0cdcafc047330f167accce9c72fcb833f0c42c2b04e"} Dec 27 09:08:35 crc kubenswrapper[4934]: I1227 09:08:35.009271 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 09:08:35 crc kubenswrapper[4934]: I1227 09:08:35.009284 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" event={"ID":"772a3e80-411a-4c15-9e1c-b702392e13af","Type":"ContainerStarted","Data":"a5ad874b224a6867308def7daedfe1277a35c413d95dda5dedcc8664ee8cf492"} Dec 27 09:08:35 crc kubenswrapper[4934]: I1227 09:08:35.372110 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-fd9565bb5-mkr88" Dec 27 09:08:36 crc kubenswrapper[4934]: I1227 09:08:36.511799 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="46b6dbac-6489-4651-8956-8c8a9767a4aa" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 09:08:36 crc kubenswrapper[4934]: I1227 09:08:36.798916 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-8msj4" Dec 27 09:08:38 crc kubenswrapper[4934]: I1227 09:08:38.467609 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:08:38 crc kubenswrapper[4934]: E1227 09:08:38.468189 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:08:41 crc kubenswrapper[4934]: I1227 09:08:41.541768 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="46b6dbac-6489-4651-8956-8c8a9767a4aa" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 09:08:46 crc kubenswrapper[4934]: I1227 09:08:46.511985 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="46b6dbac-6489-4651-8956-8c8a9767a4aa" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 09:08:50 crc kubenswrapper[4934]: I1227 09:08:50.467675 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:08:50 crc kubenswrapper[4934]: E1227 09:08:50.468443 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:08:51 crc kubenswrapper[4934]: I1227 09:08:51.521797 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="46b6dbac-6489-4651-8956-8c8a9767a4aa" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 09:08:56 crc kubenswrapper[4934]: I1227 09:08:56.516218 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="46b6dbac-6489-4651-8956-8c8a9767a4aa" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 09:09:01 crc kubenswrapper[4934]: I1227 09:09:01.513728 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="46b6dbac-6489-4651-8956-8c8a9767a4aa" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 09:09:03 crc kubenswrapper[4934]: I1227 09:09:03.358048 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/3.log" Dec 27 09:09:03 crc kubenswrapper[4934]: I1227 09:09:03.360393 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/2.log" Dec 27 09:09:03 crc kubenswrapper[4934]: I1227 09:09:03.361529 4934 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="f1d3ac602edb9810f98f35dd2e1f3dc7b5107e9c91cb4b25e87ae92dce30f7b7" exitCode=137 Dec 27 09:09:03 crc kubenswrapper[4934]: I1227 09:09:03.361568 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"f1d3ac602edb9810f98f35dd2e1f3dc7b5107e9c91cb4b25e87ae92dce30f7b7"} Dec 27 09:09:03 crc kubenswrapper[4934]: I1227 09:09:03.361600 4934 scope.go:117] "RemoveContainer" containerID="6d66e1134058289fe715318d0983996f62e222c539a7a940eb1ab6feb1889204" Dec 27 09:09:04 crc kubenswrapper[4934]: I1227 09:09:04.149211 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 27 09:09:04 crc kubenswrapper[4934]: I1227 09:09:04.376589 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/3.log" Dec 27 09:09:04 crc kubenswrapper[4934]: I1227 09:09:04.377643 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"90bbc9593c2299bc67a69dbd2d5c3fd0f0ad93d3bafbebe7455c5458c020e2d8"} Dec 27 09:09:05 crc kubenswrapper[4934]: I1227 09:09:05.469134 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:09:05 crc kubenswrapper[4934]: E1227 09:09:05.470297 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:09:06 crc kubenswrapper[4934]: I1227 09:09:06.353747 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 09:09:06 crc kubenswrapper[4934]: I1227 09:09:06.516738 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="46b6dbac-6489-4651-8956-8c8a9767a4aa" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 09:09:11 crc kubenswrapper[4934]: I1227 09:09:11.520220 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="46b6dbac-6489-4651-8956-8c8a9767a4aa" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 09:09:12 crc kubenswrapper[4934]: I1227 09:09:12.391614 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 09:09:12 crc kubenswrapper[4934]: I1227 09:09:12.398570 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 09:09:16 crc kubenswrapper[4934]: I1227 09:09:16.363625 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 27 09:09:16 crc kubenswrapper[4934]: I1227 09:09:16.467503 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:09:16 crc kubenswrapper[4934]: E1227 09:09:16.468203 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:09:16 crc kubenswrapper[4934]: I1227 09:09:16.502073 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="46b6dbac-6489-4651-8956-8c8a9767a4aa" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 09:09:21 crc kubenswrapper[4934]: I1227 09:09:21.511127 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="46b6dbac-6489-4651-8956-8c8a9767a4aa" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 09:09:26 crc kubenswrapper[4934]: I1227 09:09:26.523997 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="46b6dbac-6489-4651-8956-8c8a9767a4aa" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 27 09:09:26 crc kubenswrapper[4934]: I1227 09:09:26.524712 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 27 09:09:26 crc kubenswrapper[4934]: I1227 09:09:26.525892 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cinder-scheduler" containerStatusID={"Type":"cri-o","ID":"d26d96442fc0580edf0bce22c9b74c370ea1abc1d5b85a4a11b87445dc80c962"} pod="openstack/cinder-scheduler-0" containerMessage="Container cinder-scheduler failed startup probe, will be restarted" Dec 27 09:09:26 crc kubenswrapper[4934]: I1227 09:09:26.525952 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="46b6dbac-6489-4651-8956-8c8a9767a4aa" containerName="cinder-scheduler" containerID="cri-o://d26d96442fc0580edf0bce22c9b74c370ea1abc1d5b85a4a11b87445dc80c962" gracePeriod=30 Dec 27 09:09:26 crc kubenswrapper[4934]: I1227 09:09:26.975896 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 09:09:26 crc kubenswrapper[4934]: I1227 09:09:26.976917 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="ceilometer-notification-agent" containerID="cri-o://9a1e16b5a1faeb0c73061b916b68268113d66cfd5ed3b6d232dc6f1725c20d73" gracePeriod=30 Dec 27 09:09:26 crc kubenswrapper[4934]: I1227 09:09:26.977663 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="ceilometer-central-agent" containerID="cri-o://4b06d76681d3d075d4efcabe254bf4613f7808f693f7a1abb914d26a80a561cd" gracePeriod=30 Dec 27 09:09:26 crc kubenswrapper[4934]: I1227 09:09:26.977735 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="proxy-httpd" containerID="cri-o://836924e3f43fe5210bf012ae497e36df6e8228dd024a1dc5ca8b5c2b559b00de" gracePeriod=30 Dec 27 09:09:26 crc kubenswrapper[4934]: I1227 09:09:26.977769 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="sg-core" containerID="cri-o://ad0bb3214f6e29f8bf62f063a50e21a4b4a219cd599143f173a6e5e1b87f8532" gracePeriod=30 Dec 27 09:09:27 crc kubenswrapper[4934]: I1227 09:09:27.729879 4934 generic.go:334] "Generic (PLEG): container finished" podID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerID="4b06d76681d3d075d4efcabe254bf4613f7808f693f7a1abb914d26a80a561cd" exitCode=0 Dec 27 09:09:27 crc kubenswrapper[4934]: I1227 09:09:27.731472 4934 generic.go:334] "Generic (PLEG): container finished" podID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerID="836924e3f43fe5210bf012ae497e36df6e8228dd024a1dc5ca8b5c2b559b00de" exitCode=0 Dec 27 09:09:27 crc kubenswrapper[4934]: I1227 09:09:27.731491 4934 generic.go:334] "Generic (PLEG): container finished" podID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerID="ad0bb3214f6e29f8bf62f063a50e21a4b4a219cd599143f173a6e5e1b87f8532" exitCode=2 Dec 27 09:09:27 crc kubenswrapper[4934]: I1227 09:09:27.731517 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12e88dda-31ca-4b5b-93c3-7eced0cacda7","Type":"ContainerDied","Data":"4b06d76681d3d075d4efcabe254bf4613f7808f693f7a1abb914d26a80a561cd"} Dec 27 09:09:27 crc kubenswrapper[4934]: I1227 09:09:27.731552 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12e88dda-31ca-4b5b-93c3-7eced0cacda7","Type":"ContainerDied","Data":"836924e3f43fe5210bf012ae497e36df6e8228dd024a1dc5ca8b5c2b559b00de"} Dec 27 09:09:27 crc kubenswrapper[4934]: I1227 09:09:27.731566 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12e88dda-31ca-4b5b-93c3-7eced0cacda7","Type":"ContainerDied","Data":"ad0bb3214f6e29f8bf62f063a50e21a4b4a219cd599143f173a6e5e1b87f8532"} Dec 27 09:09:27 crc kubenswrapper[4934]: I1227 09:09:27.731584 4934 scope.go:117] "RemoveContainer" containerID="9995c35f73f77d890191ce4b613fd43a168910498934a2d034fa50716244fb04" Dec 27 09:09:29 crc kubenswrapper[4934]: I1227 09:09:29.688809 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 27 09:09:29 crc kubenswrapper[4934]: I1227 09:09:29.765721 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 27 09:09:31 crc kubenswrapper[4934]: I1227 09:09:31.360476 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 27 09:09:31 crc kubenswrapper[4934]: I1227 09:09:31.467878 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:09:31 crc kubenswrapper[4934]: E1227 09:09:31.468425 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:09:31 crc kubenswrapper[4934]: I1227 09:09:31.783627 4934 generic.go:334] "Generic (PLEG): container finished" podID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerID="9a1e16b5a1faeb0c73061b916b68268113d66cfd5ed3b6d232dc6f1725c20d73" exitCode=0 Dec 27 09:09:31 crc kubenswrapper[4934]: I1227 09:09:31.783679 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12e88dda-31ca-4b5b-93c3-7eced0cacda7","Type":"ContainerDied","Data":"9a1e16b5a1faeb0c73061b916b68268113d66cfd5ed3b6d232dc6f1725c20d73"} Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.017699 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.096703 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.212783 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-config-data\") pod \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.212939 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12e88dda-31ca-4b5b-93c3-7eced0cacda7-run-httpd\") pod \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.213033 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-ceilometer-tls-certs\") pod \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.213101 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-scripts\") pod \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.213128 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-combined-ca-bundle\") pod \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.213329 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzb6v\" (UniqueName: \"kubernetes.io/projected/12e88dda-31ca-4b5b-93c3-7eced0cacda7-kube-api-access-qzb6v\") pod \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.213422 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-sg-core-conf-yaml\") pod \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.213493 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12e88dda-31ca-4b5b-93c3-7eced0cacda7-log-httpd\") pod \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\" (UID: \"12e88dda-31ca-4b5b-93c3-7eced0cacda7\") " Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.213934 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12e88dda-31ca-4b5b-93c3-7eced0cacda7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "12e88dda-31ca-4b5b-93c3-7eced0cacda7" (UID: "12e88dda-31ca-4b5b-93c3-7eced0cacda7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.214575 4934 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12e88dda-31ca-4b5b-93c3-7eced0cacda7-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.214652 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12e88dda-31ca-4b5b-93c3-7eced0cacda7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "12e88dda-31ca-4b5b-93c3-7eced0cacda7" (UID: "12e88dda-31ca-4b5b-93c3-7eced0cacda7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.223472 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12e88dda-31ca-4b5b-93c3-7eced0cacda7-kube-api-access-qzb6v" (OuterVolumeSpecName: "kube-api-access-qzb6v") pod "12e88dda-31ca-4b5b-93c3-7eced0cacda7" (UID: "12e88dda-31ca-4b5b-93c3-7eced0cacda7"). InnerVolumeSpecName "kube-api-access-qzb6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.236251 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-scripts" (OuterVolumeSpecName: "scripts") pod "12e88dda-31ca-4b5b-93c3-7eced0cacda7" (UID: "12e88dda-31ca-4b5b-93c3-7eced0cacda7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.263437 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "12e88dda-31ca-4b5b-93c3-7eced0cacda7" (UID: "12e88dda-31ca-4b5b-93c3-7eced0cacda7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.303396 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "12e88dda-31ca-4b5b-93c3-7eced0cacda7" (UID: "12e88dda-31ca-4b5b-93c3-7eced0cacda7"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.307823 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12e88dda-31ca-4b5b-93c3-7eced0cacda7" (UID: "12e88dda-31ca-4b5b-93c3-7eced0cacda7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.316613 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzb6v\" (UniqueName: \"kubernetes.io/projected/12e88dda-31ca-4b5b-93c3-7eced0cacda7-kube-api-access-qzb6v\") on node \"crc\" DevicePath \"\"" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.316642 4934 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.316654 4934 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12e88dda-31ca-4b5b-93c3-7eced0cacda7-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.316667 4934 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.316678 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.316690 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.378712 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-config-data" (OuterVolumeSpecName: "config-data") pod "12e88dda-31ca-4b5b-93c3-7eced0cacda7" (UID: "12e88dda-31ca-4b5b-93c3-7eced0cacda7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.419346 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12e88dda-31ca-4b5b-93c3-7eced0cacda7-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.801557 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12e88dda-31ca-4b5b-93c3-7eced0cacda7","Type":"ContainerDied","Data":"91aaee4c9203052c554a3b97535ad1b215e4e586b8bc7550ffaf876d032a7982"} Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.801608 4934 scope.go:117] "RemoveContainer" containerID="4b06d76681d3d075d4efcabe254bf4613f7808f693f7a1abb914d26a80a561cd" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.801814 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.841729 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.847809 4934 scope.go:117] "RemoveContainer" containerID="836924e3f43fe5210bf012ae497e36df6e8228dd024a1dc5ca8b5c2b559b00de" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.855642 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.878328 4934 scope.go:117] "RemoveContainer" containerID="ad0bb3214f6e29f8bf62f063a50e21a4b4a219cd599143f173a6e5e1b87f8532" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.886878 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 27 09:09:32 crc kubenswrapper[4934]: E1227 09:09:32.888626 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="ceilometer-central-agent" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.888748 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="ceilometer-central-agent" Dec 27 09:09:32 crc kubenswrapper[4934]: E1227 09:09:32.888834 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73bc5c15-5c9d-49ae-bcd9-853df82fabbf" containerName="extract-utilities" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.888913 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="73bc5c15-5c9d-49ae-bcd9-853df82fabbf" containerName="extract-utilities" Dec 27 09:09:32 crc kubenswrapper[4934]: E1227 09:09:32.888937 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="sg-core" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.888943 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="sg-core" Dec 27 09:09:32 crc kubenswrapper[4934]: E1227 09:09:32.888963 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73bc5c15-5c9d-49ae-bcd9-853df82fabbf" containerName="extract-content" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.888969 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="73bc5c15-5c9d-49ae-bcd9-853df82fabbf" containerName="extract-content" Dec 27 09:09:32 crc kubenswrapper[4934]: E1227 09:09:32.888980 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="ceilometer-notification-agent" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.889255 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="ceilometer-notification-agent" Dec 27 09:09:32 crc kubenswrapper[4934]: E1227 09:09:32.889274 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73bc5c15-5c9d-49ae-bcd9-853df82fabbf" containerName="registry-server" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.889282 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="73bc5c15-5c9d-49ae-bcd9-853df82fabbf" containerName="registry-server" Dec 27 09:09:32 crc kubenswrapper[4934]: E1227 09:09:32.889293 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="proxy-httpd" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.889299 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="proxy-httpd" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.890270 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="ceilometer-central-agent" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.890552 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="sg-core" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.890576 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="ceilometer-notification-agent" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.890603 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="proxy-httpd" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.890615 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="73bc5c15-5c9d-49ae-bcd9-853df82fabbf" containerName="registry-server" Dec 27 09:09:32 crc kubenswrapper[4934]: E1227 09:09:32.891632 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="ceilometer-central-agent" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.891722 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="ceilometer-central-agent" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.892023 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" containerName="ceilometer-central-agent" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.896478 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.907772 4934 scope.go:117] "RemoveContainer" containerID="9a1e16b5a1faeb0c73061b916b68268113d66cfd5ed3b6d232dc6f1725c20d73" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.907920 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.907929 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.913776 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 27 09:09:32 crc kubenswrapper[4934]: I1227 09:09:32.922758 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.056963 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-config-data\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.057029 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxmhd\" (UniqueName: \"kubernetes.io/projected/047a93be-51e7-4e92-b18a-2f26cacd2c77-kube-api-access-vxmhd\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.057254 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/047a93be-51e7-4e92-b18a-2f26cacd2c77-run-httpd\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.057290 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.057343 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.057364 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.057443 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-scripts\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.057495 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/047a93be-51e7-4e92-b18a-2f26cacd2c77-log-httpd\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.159096 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/047a93be-51e7-4e92-b18a-2f26cacd2c77-run-httpd\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.159148 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.159180 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.159199 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.159270 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-scripts\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.159377 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/047a93be-51e7-4e92-b18a-2f26cacd2c77-log-httpd\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.159459 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-config-data\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.159484 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxmhd\" (UniqueName: \"kubernetes.io/projected/047a93be-51e7-4e92-b18a-2f26cacd2c77-kube-api-access-vxmhd\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.159914 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/047a93be-51e7-4e92-b18a-2f26cacd2c77-run-httpd\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.160450 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/047a93be-51e7-4e92-b18a-2f26cacd2c77-log-httpd\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.164263 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.164394 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.164575 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-config-data\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.165050 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-scripts\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.170092 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.178971 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxmhd\" (UniqueName: \"kubernetes.io/projected/047a93be-51e7-4e92-b18a-2f26cacd2c77-kube-api-access-vxmhd\") pod \"ceilometer-0\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.220670 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.485276 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12e88dda-31ca-4b5b-93c3-7eced0cacda7" path="/var/lib/kubelet/pods/12e88dda-31ca-4b5b-93c3-7eced0cacda7/volumes" Dec 27 09:09:33 crc kubenswrapper[4934]: I1227 09:09:33.737623 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 09:09:34 crc kubenswrapper[4934]: I1227 09:09:34.192933 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 09:09:34 crc kubenswrapper[4934]: I1227 09:09:34.827901 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"047a93be-51e7-4e92-b18a-2f26cacd2c77","Type":"ContainerStarted","Data":"ea7f04582a22c71468515026abcd0ba7ffc7db72490cd4b4c6f173abc14546b3"} Dec 27 09:09:34 crc kubenswrapper[4934]: I1227 09:09:34.828213 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"047a93be-51e7-4e92-b18a-2f26cacd2c77","Type":"ContainerStarted","Data":"670c7486bcb24f46e1bda95f3a10e1b1e605aebee6edb4222eac7cea145b7941"} Dec 27 09:09:35 crc kubenswrapper[4934]: I1227 09:09:35.843044 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"047a93be-51e7-4e92-b18a-2f26cacd2c77","Type":"ContainerStarted","Data":"3464367d1d748145b6ac75b59e2599a43376a4471fb6cb475856d480cea70df0"} Dec 27 09:09:36 crc kubenswrapper[4934]: I1227 09:09:36.870697 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"047a93be-51e7-4e92-b18a-2f26cacd2c77","Type":"ContainerStarted","Data":"d669359b94e5aee85a0567165a34d0ef17ade9a3e5477f4a7bac8b2ca619f7f0"} Dec 27 09:09:37 crc kubenswrapper[4934]: I1227 09:09:37.885898 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"047a93be-51e7-4e92-b18a-2f26cacd2c77","Type":"ContainerStarted","Data":"df4848b5ae579f0b9e8be5f522c098f0a350a8e4a8fe19372ed0c491935df655"} Dec 27 09:09:37 crc kubenswrapper[4934]: I1227 09:09:37.886563 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 27 09:09:37 crc kubenswrapper[4934]: I1227 09:09:37.886185 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="047a93be-51e7-4e92-b18a-2f26cacd2c77" containerName="ceilometer-central-agent" containerID="cri-o://ea7f04582a22c71468515026abcd0ba7ffc7db72490cd4b4c6f173abc14546b3" gracePeriod=30 Dec 27 09:09:37 crc kubenswrapper[4934]: I1227 09:09:37.886175 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="047a93be-51e7-4e92-b18a-2f26cacd2c77" containerName="sg-core" containerID="cri-o://d669359b94e5aee85a0567165a34d0ef17ade9a3e5477f4a7bac8b2ca619f7f0" gracePeriod=30 Dec 27 09:09:37 crc kubenswrapper[4934]: I1227 09:09:37.886204 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="047a93be-51e7-4e92-b18a-2f26cacd2c77" containerName="proxy-httpd" containerID="cri-o://df4848b5ae579f0b9e8be5f522c098f0a350a8e4a8fe19372ed0c491935df655" gracePeriod=30 Dec 27 09:09:37 crc kubenswrapper[4934]: I1227 09:09:37.886277 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="047a93be-51e7-4e92-b18a-2f26cacd2c77" containerName="ceilometer-notification-agent" containerID="cri-o://3464367d1d748145b6ac75b59e2599a43376a4471fb6cb475856d480cea70df0" gracePeriod=30 Dec 27 09:09:37 crc kubenswrapper[4934]: I1227 09:09:37.923934 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.899431645 podStartE2EDuration="5.923491209s" podCreationTimestamp="2025-12-27 09:09:32 +0000 UTC" firstStartedPulling="2025-12-27 09:09:34.159310684 +0000 UTC m=+5234.979751278" lastFinishedPulling="2025-12-27 09:09:37.183370248 +0000 UTC m=+5238.003810842" observedRunningTime="2025-12-27 09:09:37.906800974 +0000 UTC m=+5238.727241578" watchObservedRunningTime="2025-12-27 09:09:37.923491209 +0000 UTC m=+5238.743931813" Dec 27 09:09:38 crc kubenswrapper[4934]: I1227 09:09:38.902544 4934 generic.go:334] "Generic (PLEG): container finished" podID="047a93be-51e7-4e92-b18a-2f26cacd2c77" containerID="df4848b5ae579f0b9e8be5f522c098f0a350a8e4a8fe19372ed0c491935df655" exitCode=0 Dec 27 09:09:38 crc kubenswrapper[4934]: I1227 09:09:38.902890 4934 generic.go:334] "Generic (PLEG): container finished" podID="047a93be-51e7-4e92-b18a-2f26cacd2c77" containerID="d669359b94e5aee85a0567165a34d0ef17ade9a3e5477f4a7bac8b2ca619f7f0" exitCode=2 Dec 27 09:09:38 crc kubenswrapper[4934]: I1227 09:09:38.902899 4934 generic.go:334] "Generic (PLEG): container finished" podID="047a93be-51e7-4e92-b18a-2f26cacd2c77" containerID="3464367d1d748145b6ac75b59e2599a43376a4471fb6cb475856d480cea70df0" exitCode=0 Dec 27 09:09:38 crc kubenswrapper[4934]: I1227 09:09:38.902603 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"047a93be-51e7-4e92-b18a-2f26cacd2c77","Type":"ContainerDied","Data":"df4848b5ae579f0b9e8be5f522c098f0a350a8e4a8fe19372ed0c491935df655"} Dec 27 09:09:38 crc kubenswrapper[4934]: I1227 09:09:38.902932 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"047a93be-51e7-4e92-b18a-2f26cacd2c77","Type":"ContainerDied","Data":"d669359b94e5aee85a0567165a34d0ef17ade9a3e5477f4a7bac8b2ca619f7f0"} Dec 27 09:09:38 crc kubenswrapper[4934]: I1227 09:09:38.902944 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"047a93be-51e7-4e92-b18a-2f26cacd2c77","Type":"ContainerDied","Data":"3464367d1d748145b6ac75b59e2599a43376a4471fb6cb475856d480cea70df0"} Dec 27 09:09:45 crc kubenswrapper[4934]: I1227 09:09:45.467424 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:09:45 crc kubenswrapper[4934]: E1227 09:09:45.468348 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:09:45 crc kubenswrapper[4934]: I1227 09:09:45.992333 4934 generic.go:334] "Generic (PLEG): container finished" podID="047a93be-51e7-4e92-b18a-2f26cacd2c77" containerID="ea7f04582a22c71468515026abcd0ba7ffc7db72490cd4b4c6f173abc14546b3" exitCode=0 Dec 27 09:09:45 crc kubenswrapper[4934]: I1227 09:09:45.992666 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"047a93be-51e7-4e92-b18a-2f26cacd2c77","Type":"ContainerDied","Data":"ea7f04582a22c71468515026abcd0ba7ffc7db72490cd4b4c6f173abc14546b3"} Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.149038 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.311781 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-config-data\") pod \"047a93be-51e7-4e92-b18a-2f26cacd2c77\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.311829 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-scripts\") pod \"047a93be-51e7-4e92-b18a-2f26cacd2c77\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.311855 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-ceilometer-tls-certs\") pod \"047a93be-51e7-4e92-b18a-2f26cacd2c77\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.311891 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/047a93be-51e7-4e92-b18a-2f26cacd2c77-run-httpd\") pod \"047a93be-51e7-4e92-b18a-2f26cacd2c77\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.311957 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxmhd\" (UniqueName: \"kubernetes.io/projected/047a93be-51e7-4e92-b18a-2f26cacd2c77-kube-api-access-vxmhd\") pod \"047a93be-51e7-4e92-b18a-2f26cacd2c77\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.312045 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/047a93be-51e7-4e92-b18a-2f26cacd2c77-log-httpd\") pod \"047a93be-51e7-4e92-b18a-2f26cacd2c77\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.312132 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-combined-ca-bundle\") pod \"047a93be-51e7-4e92-b18a-2f26cacd2c77\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.312170 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-sg-core-conf-yaml\") pod \"047a93be-51e7-4e92-b18a-2f26cacd2c77\" (UID: \"047a93be-51e7-4e92-b18a-2f26cacd2c77\") " Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.312431 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/047a93be-51e7-4e92-b18a-2f26cacd2c77-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "047a93be-51e7-4e92-b18a-2f26cacd2c77" (UID: "047a93be-51e7-4e92-b18a-2f26cacd2c77"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.312552 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/047a93be-51e7-4e92-b18a-2f26cacd2c77-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "047a93be-51e7-4e92-b18a-2f26cacd2c77" (UID: "047a93be-51e7-4e92-b18a-2f26cacd2c77"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.313433 4934 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/047a93be-51e7-4e92-b18a-2f26cacd2c77-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.313454 4934 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/047a93be-51e7-4e92-b18a-2f26cacd2c77-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.319287 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-scripts" (OuterVolumeSpecName: "scripts") pod "047a93be-51e7-4e92-b18a-2f26cacd2c77" (UID: "047a93be-51e7-4e92-b18a-2f26cacd2c77"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.328531 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/047a93be-51e7-4e92-b18a-2f26cacd2c77-kube-api-access-vxmhd" (OuterVolumeSpecName: "kube-api-access-vxmhd") pod "047a93be-51e7-4e92-b18a-2f26cacd2c77" (UID: "047a93be-51e7-4e92-b18a-2f26cacd2c77"). InnerVolumeSpecName "kube-api-access-vxmhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.372329 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "047a93be-51e7-4e92-b18a-2f26cacd2c77" (UID: "047a93be-51e7-4e92-b18a-2f26cacd2c77"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.407135 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "047a93be-51e7-4e92-b18a-2f26cacd2c77" (UID: "047a93be-51e7-4e92-b18a-2f26cacd2c77"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.415908 4934 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-scripts\") on node \"crc\" DevicePath \"\"" Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.415945 4934 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.415959 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxmhd\" (UniqueName: \"kubernetes.io/projected/047a93be-51e7-4e92-b18a-2f26cacd2c77-kube-api-access-vxmhd\") on node \"crc\" DevicePath \"\"" Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.415969 4934 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.446198 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "047a93be-51e7-4e92-b18a-2f26cacd2c77" (UID: "047a93be-51e7-4e92-b18a-2f26cacd2c77"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.468228 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-config-data" (OuterVolumeSpecName: "config-data") pod "047a93be-51e7-4e92-b18a-2f26cacd2c77" (UID: "047a93be-51e7-4e92-b18a-2f26cacd2c77"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.518009 4934 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 27 09:09:46 crc kubenswrapper[4934]: I1227 09:09:46.518067 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/047a93be-51e7-4e92-b18a-2f26cacd2c77-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.008823 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"047a93be-51e7-4e92-b18a-2f26cacd2c77","Type":"ContainerDied","Data":"670c7486bcb24f46e1bda95f3a10e1b1e605aebee6edb4222eac7cea145b7941"} Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.008975 4934 scope.go:117] "RemoveContainer" containerID="df4848b5ae579f0b9e8be5f522c098f0a350a8e4a8fe19372ed0c491935df655" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.009564 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.057325 4934 scope.go:117] "RemoveContainer" containerID="d669359b94e5aee85a0567165a34d0ef17ade9a3e5477f4a7bac8b2ca619f7f0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.060152 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.069920 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.096975 4934 scope.go:117] "RemoveContainer" containerID="3464367d1d748145b6ac75b59e2599a43376a4471fb6cb475856d480cea70df0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.104116 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 27 09:09:47 crc kubenswrapper[4934]: E1227 09:09:47.104747 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="047a93be-51e7-4e92-b18a-2f26cacd2c77" containerName="ceilometer-central-agent" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.104815 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="047a93be-51e7-4e92-b18a-2f26cacd2c77" containerName="ceilometer-central-agent" Dec 27 09:09:47 crc kubenswrapper[4934]: E1227 09:09:47.104902 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="047a93be-51e7-4e92-b18a-2f26cacd2c77" containerName="proxy-httpd" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.104982 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="047a93be-51e7-4e92-b18a-2f26cacd2c77" containerName="proxy-httpd" Dec 27 09:09:47 crc kubenswrapper[4934]: E1227 09:09:47.105043 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="047a93be-51e7-4e92-b18a-2f26cacd2c77" containerName="ceilometer-notification-agent" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.105119 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="047a93be-51e7-4e92-b18a-2f26cacd2c77" containerName="ceilometer-notification-agent" Dec 27 09:09:47 crc kubenswrapper[4934]: E1227 09:09:47.105193 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="047a93be-51e7-4e92-b18a-2f26cacd2c77" containerName="sg-core" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.105250 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="047a93be-51e7-4e92-b18a-2f26cacd2c77" containerName="sg-core" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.115241 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="047a93be-51e7-4e92-b18a-2f26cacd2c77" containerName="ceilometer-notification-agent" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.115295 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="047a93be-51e7-4e92-b18a-2f26cacd2c77" containerName="sg-core" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.115335 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="047a93be-51e7-4e92-b18a-2f26cacd2c77" containerName="proxy-httpd" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.115347 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="047a93be-51e7-4e92-b18a-2f26cacd2c77" containerName="ceilometer-central-agent" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.118318 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.118433 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.123205 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.123419 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.123527 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.131715 4934 scope.go:117] "RemoveContainer" containerID="ea7f04582a22c71468515026abcd0ba7ffc7db72490cd4b4c6f173abc14546b3" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.237447 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.237533 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bswn7\" (UniqueName: \"kubernetes.io/projected/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-kube-api-access-bswn7\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.237566 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-log-httpd\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.237611 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.237752 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-config-data\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.238145 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-scripts\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.238184 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-run-httpd\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.238227 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.339970 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.340044 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-config-data\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.340237 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-scripts\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.340261 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-run-httpd\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.340298 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.340493 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.340579 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bswn7\" (UniqueName: \"kubernetes.io/projected/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-kube-api-access-bswn7\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.340619 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-log-httpd\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.341190 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-log-httpd\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.341931 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-run-httpd\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.345519 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-config-data\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.353447 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-scripts\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.353534 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.355797 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.356579 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bswn7\" (UniqueName: \"kubernetes.io/projected/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-kube-api-access-bswn7\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.364004 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3c21d8f2-980c-4491-b91c-a14d5c7eb20f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3c21d8f2-980c-4491-b91c-a14d5c7eb20f\") " pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.443244 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.479980 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="047a93be-51e7-4e92-b18a-2f26cacd2c77" path="/var/lib/kubelet/pods/047a93be-51e7-4e92-b18a-2f26cacd2c77/volumes" Dec 27 09:09:47 crc kubenswrapper[4934]: I1227 09:09:47.948344 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 27 09:09:48 crc kubenswrapper[4934]: I1227 09:09:48.022904 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c21d8f2-980c-4491-b91c-a14d5c7eb20f","Type":"ContainerStarted","Data":"3a4be2d28781b01989efbf82b1587e98984c94aae4936379799136db998711bd"} Dec 27 09:09:49 crc kubenswrapper[4934]: I1227 09:09:49.052111 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c21d8f2-980c-4491-b91c-a14d5c7eb20f","Type":"ContainerStarted","Data":"c3e2b1df992e74c293b882b7c2609a047e51deaa3240315fe74838f984e65dd9"} Dec 27 09:09:50 crc kubenswrapper[4934]: I1227 09:09:50.070810 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c21d8f2-980c-4491-b91c-a14d5c7eb20f","Type":"ContainerStarted","Data":"5bb72cf0cf03bba1c93be2d5862ff44ad63fd96631625abd0c8176da6e352cfa"} Dec 27 09:09:50 crc kubenswrapper[4934]: I1227 09:09:50.071431 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c21d8f2-980c-4491-b91c-a14d5c7eb20f","Type":"ContainerStarted","Data":"8e6d44bc3ee40bceb39607c332efbff2ddd39a1bb76f8a6b3ce57f68db206150"} Dec 27 09:09:52 crc kubenswrapper[4934]: I1227 09:09:52.099156 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3c21d8f2-980c-4491-b91c-a14d5c7eb20f","Type":"ContainerStarted","Data":"25be269d48801b142c8b4d471092ea94ff067c6f7256f822ba1890c790a24cee"} Dec 27 09:09:52 crc kubenswrapper[4934]: I1227 09:09:52.099818 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 27 09:09:52 crc kubenswrapper[4934]: I1227 09:09:52.127598 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.8615327449999999 podStartE2EDuration="5.127582222s" podCreationTimestamp="2025-12-27 09:09:47 +0000 UTC" firstStartedPulling="2025-12-27 09:09:47.945154224 +0000 UTC m=+5248.765594828" lastFinishedPulling="2025-12-27 09:09:51.211203711 +0000 UTC m=+5252.031644305" observedRunningTime="2025-12-27 09:09:52.123846449 +0000 UTC m=+5252.944287083" watchObservedRunningTime="2025-12-27 09:09:52.127582222 +0000 UTC m=+5252.948022816" Dec 27 09:09:56 crc kubenswrapper[4934]: I1227 09:09:56.468294 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:09:56 crc kubenswrapper[4934]: E1227 09:09:56.468904 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:09:57 crc kubenswrapper[4934]: I1227 09:09:57.161613 4934 generic.go:334] "Generic (PLEG): container finished" podID="46b6dbac-6489-4651-8956-8c8a9767a4aa" containerID="d26d96442fc0580edf0bce22c9b74c370ea1abc1d5b85a4a11b87445dc80c962" exitCode=137 Dec 27 09:09:57 crc kubenswrapper[4934]: I1227 09:09:57.161706 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"46b6dbac-6489-4651-8956-8c8a9767a4aa","Type":"ContainerDied","Data":"d26d96442fc0580edf0bce22c9b74c370ea1abc1d5b85a4a11b87445dc80c962"} Dec 27 09:09:57 crc kubenswrapper[4934]: I1227 09:09:57.161943 4934 scope.go:117] "RemoveContainer" containerID="d56b14937b8f4e2616b25041792887ddd2d75019777c58f0718b35781615cd6b" Dec 27 09:09:58 crc kubenswrapper[4934]: I1227 09:09:58.181308 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"46b6dbac-6489-4651-8956-8c8a9767a4aa","Type":"ContainerStarted","Data":"7bd343296be855e0fd277feebdbd547add4b4b8ab5c4500a211b4c8b831b7fbb"} Dec 27 09:10:01 crc kubenswrapper[4934]: I1227 09:10:01.490226 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 27 09:10:06 crc kubenswrapper[4934]: I1227 09:10:06.512574 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 27 09:10:07 crc kubenswrapper[4934]: I1227 09:10:07.467475 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:10:07 crc kubenswrapper[4934]: E1227 09:10:07.468211 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:10:17 crc kubenswrapper[4934]: I1227 09:10:17.459456 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 27 09:10:18 crc kubenswrapper[4934]: I1227 09:10:18.467598 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:10:18 crc kubenswrapper[4934]: E1227 09:10:18.467853 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:10:31 crc kubenswrapper[4934]: I1227 09:10:31.468304 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:10:31 crc kubenswrapper[4934]: E1227 09:10:31.469297 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:10:42 crc kubenswrapper[4934]: I1227 09:10:42.468660 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:10:42 crc kubenswrapper[4934]: E1227 09:10:42.469900 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:10:53 crc kubenswrapper[4934]: I1227 09:10:53.467133 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:10:53 crc kubenswrapper[4934]: E1227 09:10:53.467843 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:11:04 crc kubenswrapper[4934]: I1227 09:11:04.468125 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:11:04 crc kubenswrapper[4934]: E1227 09:11:04.468924 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:11:18 crc kubenswrapper[4934]: I1227 09:11:18.469133 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:11:19 crc kubenswrapper[4934]: I1227 09:11:19.300972 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"50b97ec86a4c85f552c63fe4f7d1063d9fc24b329031aa7650e1689dc131df87"} Dec 27 09:13:45 crc kubenswrapper[4934]: I1227 09:13:45.330853 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 09:13:45 crc kubenswrapper[4934]: I1227 09:13:45.331840 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 09:14:15 crc kubenswrapper[4934]: I1227 09:14:15.330288 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 09:14:15 crc kubenswrapper[4934]: I1227 09:14:15.331049 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 09:14:45 crc kubenswrapper[4934]: I1227 09:14:45.330679 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 09:14:45 crc kubenswrapper[4934]: I1227 09:14:45.331426 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 09:14:45 crc kubenswrapper[4934]: I1227 09:14:45.331499 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 09:14:45 crc kubenswrapper[4934]: I1227 09:14:45.332826 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"50b97ec86a4c85f552c63fe4f7d1063d9fc24b329031aa7650e1689dc131df87"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 09:14:45 crc kubenswrapper[4934]: I1227 09:14:45.332932 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://50b97ec86a4c85f552c63fe4f7d1063d9fc24b329031aa7650e1689dc131df87" gracePeriod=600 Dec 27 09:14:46 crc kubenswrapper[4934]: I1227 09:14:46.485018 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="50b97ec86a4c85f552c63fe4f7d1063d9fc24b329031aa7650e1689dc131df87" exitCode=0 Dec 27 09:14:46 crc kubenswrapper[4934]: I1227 09:14:46.485093 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"50b97ec86a4c85f552c63fe4f7d1063d9fc24b329031aa7650e1689dc131df87"} Dec 27 09:14:46 crc kubenswrapper[4934]: I1227 09:14:46.486643 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c"} Dec 27 09:14:46 crc kubenswrapper[4934]: I1227 09:14:46.486675 4934 scope.go:117] "RemoveContainer" containerID="05b54def72456fb8b59919c4fd3c5bea6da0b5a5168391a0c67731d9791690da" Dec 27 09:15:00 crc kubenswrapper[4934]: I1227 09:15:00.209863 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447115-wh5qc"] Dec 27 09:15:00 crc kubenswrapper[4934]: I1227 09:15:00.212203 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447115-wh5qc" Dec 27 09:15:00 crc kubenswrapper[4934]: I1227 09:15:00.214906 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 27 09:15:00 crc kubenswrapper[4934]: I1227 09:15:00.215788 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 27 09:15:00 crc kubenswrapper[4934]: I1227 09:15:00.222031 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447115-wh5qc"] Dec 27 09:15:00 crc kubenswrapper[4934]: I1227 09:15:00.289718 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbtqm\" (UniqueName: \"kubernetes.io/projected/02fb2162-3fc2-4349-9633-56aabf11af23-kube-api-access-gbtqm\") pod \"collect-profiles-29447115-wh5qc\" (UID: \"02fb2162-3fc2-4349-9633-56aabf11af23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447115-wh5qc" Dec 27 09:15:00 crc kubenswrapper[4934]: I1227 09:15:00.289804 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/02fb2162-3fc2-4349-9633-56aabf11af23-config-volume\") pod \"collect-profiles-29447115-wh5qc\" (UID: \"02fb2162-3fc2-4349-9633-56aabf11af23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447115-wh5qc" Dec 27 09:15:00 crc kubenswrapper[4934]: I1227 09:15:00.289839 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/02fb2162-3fc2-4349-9633-56aabf11af23-secret-volume\") pod \"collect-profiles-29447115-wh5qc\" (UID: \"02fb2162-3fc2-4349-9633-56aabf11af23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447115-wh5qc" Dec 27 09:15:00 crc kubenswrapper[4934]: I1227 09:15:00.392561 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/02fb2162-3fc2-4349-9633-56aabf11af23-config-volume\") pod \"collect-profiles-29447115-wh5qc\" (UID: \"02fb2162-3fc2-4349-9633-56aabf11af23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447115-wh5qc" Dec 27 09:15:00 crc kubenswrapper[4934]: I1227 09:15:00.392612 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/02fb2162-3fc2-4349-9633-56aabf11af23-secret-volume\") pod \"collect-profiles-29447115-wh5qc\" (UID: \"02fb2162-3fc2-4349-9633-56aabf11af23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447115-wh5qc" Dec 27 09:15:00 crc kubenswrapper[4934]: I1227 09:15:00.392781 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbtqm\" (UniqueName: \"kubernetes.io/projected/02fb2162-3fc2-4349-9633-56aabf11af23-kube-api-access-gbtqm\") pod \"collect-profiles-29447115-wh5qc\" (UID: \"02fb2162-3fc2-4349-9633-56aabf11af23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447115-wh5qc" Dec 27 09:15:00 crc kubenswrapper[4934]: I1227 09:15:00.394516 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/02fb2162-3fc2-4349-9633-56aabf11af23-config-volume\") pod \"collect-profiles-29447115-wh5qc\" (UID: \"02fb2162-3fc2-4349-9633-56aabf11af23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447115-wh5qc" Dec 27 09:15:01 crc kubenswrapper[4934]: I1227 09:15:01.140802 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/02fb2162-3fc2-4349-9633-56aabf11af23-secret-volume\") pod \"collect-profiles-29447115-wh5qc\" (UID: \"02fb2162-3fc2-4349-9633-56aabf11af23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447115-wh5qc" Dec 27 09:15:01 crc kubenswrapper[4934]: I1227 09:15:01.148126 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbtqm\" (UniqueName: \"kubernetes.io/projected/02fb2162-3fc2-4349-9633-56aabf11af23-kube-api-access-gbtqm\") pod \"collect-profiles-29447115-wh5qc\" (UID: \"02fb2162-3fc2-4349-9633-56aabf11af23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447115-wh5qc" Dec 27 09:15:01 crc kubenswrapper[4934]: I1227 09:15:01.435919 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447115-wh5qc" Dec 27 09:15:01 crc kubenswrapper[4934]: I1227 09:15:01.934602 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447115-wh5qc"] Dec 27 09:15:02 crc kubenswrapper[4934]: I1227 09:15:02.723359 4934 generic.go:334] "Generic (PLEG): container finished" podID="02fb2162-3fc2-4349-9633-56aabf11af23" containerID="0c3d80bc8765b3c55f5b4f1534262c9dae632d84da71a772cb3c2af5fafc458c" exitCode=0 Dec 27 09:15:02 crc kubenswrapper[4934]: I1227 09:15:02.723709 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447115-wh5qc" event={"ID":"02fb2162-3fc2-4349-9633-56aabf11af23","Type":"ContainerDied","Data":"0c3d80bc8765b3c55f5b4f1534262c9dae632d84da71a772cb3c2af5fafc458c"} Dec 27 09:15:02 crc kubenswrapper[4934]: I1227 09:15:02.724702 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447115-wh5qc" event={"ID":"02fb2162-3fc2-4349-9633-56aabf11af23","Type":"ContainerStarted","Data":"37c9064c1daa07ea68aee4ae7de3e4a9f2ca091d0dc6858c6ebdabe45f604057"} Dec 27 09:15:04 crc kubenswrapper[4934]: I1227 09:15:04.227747 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447115-wh5qc" Dec 27 09:15:04 crc kubenswrapper[4934]: I1227 09:15:04.291357 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/02fb2162-3fc2-4349-9633-56aabf11af23-config-volume\") pod \"02fb2162-3fc2-4349-9633-56aabf11af23\" (UID: \"02fb2162-3fc2-4349-9633-56aabf11af23\") " Dec 27 09:15:04 crc kubenswrapper[4934]: I1227 09:15:04.291619 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbtqm\" (UniqueName: \"kubernetes.io/projected/02fb2162-3fc2-4349-9633-56aabf11af23-kube-api-access-gbtqm\") pod \"02fb2162-3fc2-4349-9633-56aabf11af23\" (UID: \"02fb2162-3fc2-4349-9633-56aabf11af23\") " Dec 27 09:15:04 crc kubenswrapper[4934]: I1227 09:15:04.291687 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/02fb2162-3fc2-4349-9633-56aabf11af23-secret-volume\") pod \"02fb2162-3fc2-4349-9633-56aabf11af23\" (UID: \"02fb2162-3fc2-4349-9633-56aabf11af23\") " Dec 27 09:15:04 crc kubenswrapper[4934]: I1227 09:15:04.292500 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02fb2162-3fc2-4349-9633-56aabf11af23-config-volume" (OuterVolumeSpecName: "config-volume") pod "02fb2162-3fc2-4349-9633-56aabf11af23" (UID: "02fb2162-3fc2-4349-9633-56aabf11af23"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 09:15:04 crc kubenswrapper[4934]: I1227 09:15:04.299718 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02fb2162-3fc2-4349-9633-56aabf11af23-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "02fb2162-3fc2-4349-9633-56aabf11af23" (UID: "02fb2162-3fc2-4349-9633-56aabf11af23"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 09:15:04 crc kubenswrapper[4934]: I1227 09:15:04.304146 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02fb2162-3fc2-4349-9633-56aabf11af23-kube-api-access-gbtqm" (OuterVolumeSpecName: "kube-api-access-gbtqm") pod "02fb2162-3fc2-4349-9633-56aabf11af23" (UID: "02fb2162-3fc2-4349-9633-56aabf11af23"). InnerVolumeSpecName "kube-api-access-gbtqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:15:04 crc kubenswrapper[4934]: I1227 09:15:04.395055 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbtqm\" (UniqueName: \"kubernetes.io/projected/02fb2162-3fc2-4349-9633-56aabf11af23-kube-api-access-gbtqm\") on node \"crc\" DevicePath \"\"" Dec 27 09:15:04 crc kubenswrapper[4934]: I1227 09:15:04.395105 4934 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/02fb2162-3fc2-4349-9633-56aabf11af23-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 27 09:15:04 crc kubenswrapper[4934]: I1227 09:15:04.395116 4934 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/02fb2162-3fc2-4349-9633-56aabf11af23-config-volume\") on node \"crc\" DevicePath \"\"" Dec 27 09:15:04 crc kubenswrapper[4934]: I1227 09:15:04.748466 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447115-wh5qc" Dec 27 09:15:04 crc kubenswrapper[4934]: I1227 09:15:04.748359 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447115-wh5qc" event={"ID":"02fb2162-3fc2-4349-9633-56aabf11af23","Type":"ContainerDied","Data":"37c9064c1daa07ea68aee4ae7de3e4a9f2ca091d0dc6858c6ebdabe45f604057"} Dec 27 09:15:04 crc kubenswrapper[4934]: I1227 09:15:04.749221 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37c9064c1daa07ea68aee4ae7de3e4a9f2ca091d0dc6858c6ebdabe45f604057" Dec 27 09:15:05 crc kubenswrapper[4934]: I1227 09:15:05.314937 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl"] Dec 27 09:15:05 crc kubenswrapper[4934]: I1227 09:15:05.328516 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447070-hwzwl"] Dec 27 09:15:05 crc kubenswrapper[4934]: I1227 09:15:05.488739 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f809b83-b833-46e0-9ff7-d52c567711e9" path="/var/lib/kubelet/pods/2f809b83-b833-46e0-9ff7-d52c567711e9/volumes" Dec 27 09:15:30 crc kubenswrapper[4934]: I1227 09:15:30.892093 4934 scope.go:117] "RemoveContainer" containerID="ace9f6c99252fcb22eb3a5770f78a4d70b8b8b2e3168809a11c3e4346e78edc6" Dec 27 09:16:00 crc kubenswrapper[4934]: I1227 09:16:00.459198 4934 trace.go:236] Trace[1845133649]: "Calculate volume metrics of ovndbcluster-nb-etc-ovn for pod openstack/ovsdbserver-nb-0" (27-Dec-2025 09:15:59.400) (total time: 1058ms): Dec 27 09:16:00 crc kubenswrapper[4934]: Trace[1845133649]: [1.058470212s] [1.058470212s] END Dec 27 09:16:45 crc kubenswrapper[4934]: I1227 09:16:45.330878 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 09:16:45 crc kubenswrapper[4934]: I1227 09:16:45.331558 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 09:17:15 crc kubenswrapper[4934]: I1227 09:17:15.329785 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 09:17:15 crc kubenswrapper[4934]: I1227 09:17:15.330423 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 09:17:36 crc kubenswrapper[4934]: I1227 09:17:36.295771 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2sc7g"] Dec 27 09:17:36 crc kubenswrapper[4934]: E1227 09:17:36.296814 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02fb2162-3fc2-4349-9633-56aabf11af23" containerName="collect-profiles" Dec 27 09:17:36 crc kubenswrapper[4934]: I1227 09:17:36.296829 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="02fb2162-3fc2-4349-9633-56aabf11af23" containerName="collect-profiles" Dec 27 09:17:36 crc kubenswrapper[4934]: I1227 09:17:36.297096 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="02fb2162-3fc2-4349-9633-56aabf11af23" containerName="collect-profiles" Dec 27 09:17:36 crc kubenswrapper[4934]: I1227 09:17:36.300674 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2sc7g" Dec 27 09:17:36 crc kubenswrapper[4934]: I1227 09:17:36.314385 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2sc7g"] Dec 27 09:17:36 crc kubenswrapper[4934]: I1227 09:17:36.331244 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6e90da6-e926-4a6a-8a61-a418b5cd15d7-utilities\") pod \"certified-operators-2sc7g\" (UID: \"d6e90da6-e926-4a6a-8a61-a418b5cd15d7\") " pod="openshift-marketplace/certified-operators-2sc7g" Dec 27 09:17:36 crc kubenswrapper[4934]: I1227 09:17:36.331354 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl75s\" (UniqueName: \"kubernetes.io/projected/d6e90da6-e926-4a6a-8a61-a418b5cd15d7-kube-api-access-dl75s\") pod \"certified-operators-2sc7g\" (UID: \"d6e90da6-e926-4a6a-8a61-a418b5cd15d7\") " pod="openshift-marketplace/certified-operators-2sc7g" Dec 27 09:17:36 crc kubenswrapper[4934]: I1227 09:17:36.331840 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6e90da6-e926-4a6a-8a61-a418b5cd15d7-catalog-content\") pod \"certified-operators-2sc7g\" (UID: \"d6e90da6-e926-4a6a-8a61-a418b5cd15d7\") " pod="openshift-marketplace/certified-operators-2sc7g" Dec 27 09:17:36 crc kubenswrapper[4934]: I1227 09:17:36.434280 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6e90da6-e926-4a6a-8a61-a418b5cd15d7-utilities\") pod \"certified-operators-2sc7g\" (UID: \"d6e90da6-e926-4a6a-8a61-a418b5cd15d7\") " pod="openshift-marketplace/certified-operators-2sc7g" Dec 27 09:17:36 crc kubenswrapper[4934]: I1227 09:17:36.434969 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6e90da6-e926-4a6a-8a61-a418b5cd15d7-utilities\") pod \"certified-operators-2sc7g\" (UID: \"d6e90da6-e926-4a6a-8a61-a418b5cd15d7\") " pod="openshift-marketplace/certified-operators-2sc7g" Dec 27 09:17:36 crc kubenswrapper[4934]: I1227 09:17:36.435042 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl75s\" (UniqueName: \"kubernetes.io/projected/d6e90da6-e926-4a6a-8a61-a418b5cd15d7-kube-api-access-dl75s\") pod \"certified-operators-2sc7g\" (UID: \"d6e90da6-e926-4a6a-8a61-a418b5cd15d7\") " pod="openshift-marketplace/certified-operators-2sc7g" Dec 27 09:17:36 crc kubenswrapper[4934]: I1227 09:17:36.435262 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6e90da6-e926-4a6a-8a61-a418b5cd15d7-catalog-content\") pod \"certified-operators-2sc7g\" (UID: \"d6e90da6-e926-4a6a-8a61-a418b5cd15d7\") " pod="openshift-marketplace/certified-operators-2sc7g" Dec 27 09:17:36 crc kubenswrapper[4934]: I1227 09:17:36.435973 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6e90da6-e926-4a6a-8a61-a418b5cd15d7-catalog-content\") pod \"certified-operators-2sc7g\" (UID: \"d6e90da6-e926-4a6a-8a61-a418b5cd15d7\") " pod="openshift-marketplace/certified-operators-2sc7g" Dec 27 09:17:36 crc kubenswrapper[4934]: I1227 09:17:36.457854 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl75s\" (UniqueName: \"kubernetes.io/projected/d6e90da6-e926-4a6a-8a61-a418b5cd15d7-kube-api-access-dl75s\") pod \"certified-operators-2sc7g\" (UID: \"d6e90da6-e926-4a6a-8a61-a418b5cd15d7\") " pod="openshift-marketplace/certified-operators-2sc7g" Dec 27 09:17:36 crc kubenswrapper[4934]: I1227 09:17:36.627584 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2sc7g" Dec 27 09:17:37 crc kubenswrapper[4934]: I1227 09:17:37.149676 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2sc7g"] Dec 27 09:17:37 crc kubenswrapper[4934]: I1227 09:17:37.742913 4934 generic.go:334] "Generic (PLEG): container finished" podID="d6e90da6-e926-4a6a-8a61-a418b5cd15d7" containerID="5d2fcb8d4c013561b4b3cfe12e0a97dc8f28694d8fdf5f9ea61a2eccd224ffe8" exitCode=0 Dec 27 09:17:37 crc kubenswrapper[4934]: I1227 09:17:37.742985 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2sc7g" event={"ID":"d6e90da6-e926-4a6a-8a61-a418b5cd15d7","Type":"ContainerDied","Data":"5d2fcb8d4c013561b4b3cfe12e0a97dc8f28694d8fdf5f9ea61a2eccd224ffe8"} Dec 27 09:17:37 crc kubenswrapper[4934]: I1227 09:17:37.743281 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2sc7g" event={"ID":"d6e90da6-e926-4a6a-8a61-a418b5cd15d7","Type":"ContainerStarted","Data":"1dd0eb9520faf01e7de43da8093771ccb32e6c02b8b8c4753ee308e4df4f94c5"} Dec 27 09:17:37 crc kubenswrapper[4934]: I1227 09:17:37.749050 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 27 09:17:39 crc kubenswrapper[4934]: I1227 09:17:39.772176 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2sc7g" event={"ID":"d6e90da6-e926-4a6a-8a61-a418b5cd15d7","Type":"ContainerStarted","Data":"bc4c1de8c2f2f32872a107e4d05a4814cd0766b49d405a1ac8e949473712f430"} Dec 27 09:17:42 crc kubenswrapper[4934]: I1227 09:17:42.221051 4934 generic.go:334] "Generic (PLEG): container finished" podID="d6e90da6-e926-4a6a-8a61-a418b5cd15d7" containerID="bc4c1de8c2f2f32872a107e4d05a4814cd0766b49d405a1ac8e949473712f430" exitCode=0 Dec 27 09:17:42 crc kubenswrapper[4934]: I1227 09:17:42.256006 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2sc7g" event={"ID":"d6e90da6-e926-4a6a-8a61-a418b5cd15d7","Type":"ContainerDied","Data":"bc4c1de8c2f2f32872a107e4d05a4814cd0766b49d405a1ac8e949473712f430"} Dec 27 09:17:43 crc kubenswrapper[4934]: I1227 09:17:43.238147 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2sc7g" event={"ID":"d6e90da6-e926-4a6a-8a61-a418b5cd15d7","Type":"ContainerStarted","Data":"450bb0fa53d0f90dfa7edb99cac4e09c3c4a60ceff0584d3d3e446aeea40e077"} Dec 27 09:17:43 crc kubenswrapper[4934]: I1227 09:17:43.264358 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2sc7g" podStartSLOduration=2.36041071 podStartE2EDuration="7.264333605s" podCreationTimestamp="2025-12-27 09:17:36 +0000 UTC" firstStartedPulling="2025-12-27 09:17:37.747574024 +0000 UTC m=+5718.568014628" lastFinishedPulling="2025-12-27 09:17:42.651496929 +0000 UTC m=+5723.471937523" observedRunningTime="2025-12-27 09:17:43.255106556 +0000 UTC m=+5724.075547160" watchObservedRunningTime="2025-12-27 09:17:43.264333605 +0000 UTC m=+5724.084774199" Dec 27 09:17:45 crc kubenswrapper[4934]: I1227 09:17:45.329953 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 09:17:45 crc kubenswrapper[4934]: I1227 09:17:45.330306 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 09:17:45 crc kubenswrapper[4934]: I1227 09:17:45.330359 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 09:17:45 crc kubenswrapper[4934]: I1227 09:17:45.331566 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 09:17:45 crc kubenswrapper[4934]: I1227 09:17:45.331850 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" gracePeriod=600 Dec 27 09:17:46 crc kubenswrapper[4934]: I1227 09:17:46.273147 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" exitCode=0 Dec 27 09:17:46 crc kubenswrapper[4934]: I1227 09:17:46.273202 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c"} Dec 27 09:17:46 crc kubenswrapper[4934]: I1227 09:17:46.273534 4934 scope.go:117] "RemoveContainer" containerID="50b97ec86a4c85f552c63fe4f7d1063d9fc24b329031aa7650e1689dc131df87" Dec 27 09:17:46 crc kubenswrapper[4934]: E1227 09:17:46.369234 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:17:46 crc kubenswrapper[4934]: I1227 09:17:46.628601 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2sc7g" Dec 27 09:17:46 crc kubenswrapper[4934]: I1227 09:17:46.628668 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2sc7g" Dec 27 09:17:47 crc kubenswrapper[4934]: I1227 09:17:47.289646 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:17:47 crc kubenswrapper[4934]: E1227 09:17:47.290241 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:17:47 crc kubenswrapper[4934]: I1227 09:17:47.683732 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-2sc7g" podUID="d6e90da6-e926-4a6a-8a61-a418b5cd15d7" containerName="registry-server" probeResult="failure" output=< Dec 27 09:17:47 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 09:17:47 crc kubenswrapper[4934]: > Dec 27 09:17:51 crc kubenswrapper[4934]: I1227 09:17:51.624572 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hzlkj"] Dec 27 09:17:51 crc kubenswrapper[4934]: I1227 09:17:51.650297 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hzlkj" Dec 27 09:17:51 crc kubenswrapper[4934]: I1227 09:17:51.664668 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hzlkj"] Dec 27 09:17:51 crc kubenswrapper[4934]: I1227 09:17:51.807925 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f972030f-cb58-420b-927e-80ea4ccb198b-utilities\") pod \"community-operators-hzlkj\" (UID: \"f972030f-cb58-420b-927e-80ea4ccb198b\") " pod="openshift-marketplace/community-operators-hzlkj" Dec 27 09:17:51 crc kubenswrapper[4934]: I1227 09:17:51.808131 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f972030f-cb58-420b-927e-80ea4ccb198b-catalog-content\") pod \"community-operators-hzlkj\" (UID: \"f972030f-cb58-420b-927e-80ea4ccb198b\") " pod="openshift-marketplace/community-operators-hzlkj" Dec 27 09:17:51 crc kubenswrapper[4934]: I1227 09:17:51.808167 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv85q\" (UniqueName: \"kubernetes.io/projected/f972030f-cb58-420b-927e-80ea4ccb198b-kube-api-access-xv85q\") pod \"community-operators-hzlkj\" (UID: \"f972030f-cb58-420b-927e-80ea4ccb198b\") " pod="openshift-marketplace/community-operators-hzlkj" Dec 27 09:17:51 crc kubenswrapper[4934]: I1227 09:17:51.910997 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f972030f-cb58-420b-927e-80ea4ccb198b-catalog-content\") pod \"community-operators-hzlkj\" (UID: \"f972030f-cb58-420b-927e-80ea4ccb198b\") " pod="openshift-marketplace/community-operators-hzlkj" Dec 27 09:17:51 crc kubenswrapper[4934]: I1227 09:17:51.911514 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv85q\" (UniqueName: \"kubernetes.io/projected/f972030f-cb58-420b-927e-80ea4ccb198b-kube-api-access-xv85q\") pod \"community-operators-hzlkj\" (UID: \"f972030f-cb58-420b-927e-80ea4ccb198b\") " pod="openshift-marketplace/community-operators-hzlkj" Dec 27 09:17:51 crc kubenswrapper[4934]: I1227 09:17:51.911721 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f972030f-cb58-420b-927e-80ea4ccb198b-catalog-content\") pod \"community-operators-hzlkj\" (UID: \"f972030f-cb58-420b-927e-80ea4ccb198b\") " pod="openshift-marketplace/community-operators-hzlkj" Dec 27 09:17:51 crc kubenswrapper[4934]: I1227 09:17:51.912065 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f972030f-cb58-420b-927e-80ea4ccb198b-utilities\") pod \"community-operators-hzlkj\" (UID: \"f972030f-cb58-420b-927e-80ea4ccb198b\") " pod="openshift-marketplace/community-operators-hzlkj" Dec 27 09:17:51 crc kubenswrapper[4934]: I1227 09:17:51.912527 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f972030f-cb58-420b-927e-80ea4ccb198b-utilities\") pod \"community-operators-hzlkj\" (UID: \"f972030f-cb58-420b-927e-80ea4ccb198b\") " pod="openshift-marketplace/community-operators-hzlkj" Dec 27 09:17:51 crc kubenswrapper[4934]: I1227 09:17:51.932875 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv85q\" (UniqueName: \"kubernetes.io/projected/f972030f-cb58-420b-927e-80ea4ccb198b-kube-api-access-xv85q\") pod \"community-operators-hzlkj\" (UID: \"f972030f-cb58-420b-927e-80ea4ccb198b\") " pod="openshift-marketplace/community-operators-hzlkj" Dec 27 09:17:51 crc kubenswrapper[4934]: I1227 09:17:51.982452 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hzlkj" Dec 27 09:17:52 crc kubenswrapper[4934]: I1227 09:17:52.537349 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hzlkj"] Dec 27 09:17:53 crc kubenswrapper[4934]: I1227 09:17:53.362766 4934 generic.go:334] "Generic (PLEG): container finished" podID="f972030f-cb58-420b-927e-80ea4ccb198b" containerID="72b880a5187c780e9810bba436455bfd3701aabe5aa7eb3b00ce71ea1711f3cb" exitCode=0 Dec 27 09:17:53 crc kubenswrapper[4934]: I1227 09:17:53.362880 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzlkj" event={"ID":"f972030f-cb58-420b-927e-80ea4ccb198b","Type":"ContainerDied","Data":"72b880a5187c780e9810bba436455bfd3701aabe5aa7eb3b00ce71ea1711f3cb"} Dec 27 09:17:53 crc kubenswrapper[4934]: I1227 09:17:53.363409 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzlkj" event={"ID":"f972030f-cb58-420b-927e-80ea4ccb198b","Type":"ContainerStarted","Data":"0fdc9e4e804491916b3f8fba768e9c13db530c2cd7dc7d3aca95dcb8624e7e10"} Dec 27 09:17:55 crc kubenswrapper[4934]: I1227 09:17:55.390716 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzlkj" event={"ID":"f972030f-cb58-420b-927e-80ea4ccb198b","Type":"ContainerStarted","Data":"77d33799b02558dd1e41317bbf461aee0ad5d7d0780c0e3ec14583fdb19fe23b"} Dec 27 09:17:56 crc kubenswrapper[4934]: I1227 09:17:56.405125 4934 generic.go:334] "Generic (PLEG): container finished" podID="f972030f-cb58-420b-927e-80ea4ccb198b" containerID="77d33799b02558dd1e41317bbf461aee0ad5d7d0780c0e3ec14583fdb19fe23b" exitCode=0 Dec 27 09:17:56 crc kubenswrapper[4934]: I1227 09:17:56.405199 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzlkj" event={"ID":"f972030f-cb58-420b-927e-80ea4ccb198b","Type":"ContainerDied","Data":"77d33799b02558dd1e41317bbf461aee0ad5d7d0780c0e3ec14583fdb19fe23b"} Dec 27 09:17:56 crc kubenswrapper[4934]: I1227 09:17:56.683361 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2sc7g" Dec 27 09:17:56 crc kubenswrapper[4934]: I1227 09:17:56.739254 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2sc7g" Dec 27 09:17:57 crc kubenswrapper[4934]: I1227 09:17:57.423271 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzlkj" event={"ID":"f972030f-cb58-420b-927e-80ea4ccb198b","Type":"ContainerStarted","Data":"7ef2085cf460dc4930f4658c8290f48e38cc30e41bd78760b12ecee5605fbb3b"} Dec 27 09:17:57 crc kubenswrapper[4934]: I1227 09:17:57.461354 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hzlkj" podStartSLOduration=3.001952021 podStartE2EDuration="6.461327568s" podCreationTimestamp="2025-12-27 09:17:51 +0000 UTC" firstStartedPulling="2025-12-27 09:17:53.365681446 +0000 UTC m=+5734.186122040" lastFinishedPulling="2025-12-27 09:17:56.825056993 +0000 UTC m=+5737.645497587" observedRunningTime="2025-12-27 09:17:57.448273055 +0000 UTC m=+5738.268713679" watchObservedRunningTime="2025-12-27 09:17:57.461327568 +0000 UTC m=+5738.281768182" Dec 27 09:17:58 crc kubenswrapper[4934]: I1227 09:17:58.388451 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2sc7g"] Dec 27 09:17:58 crc kubenswrapper[4934]: I1227 09:17:58.433944 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2sc7g" podUID="d6e90da6-e926-4a6a-8a61-a418b5cd15d7" containerName="registry-server" containerID="cri-o://450bb0fa53d0f90dfa7edb99cac4e09c3c4a60ceff0584d3d3e446aeea40e077" gracePeriod=2 Dec 27 09:17:58 crc kubenswrapper[4934]: I1227 09:17:58.468283 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:17:58 crc kubenswrapper[4934]: E1227 09:17:58.468854 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.014751 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2sc7g" Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.085272 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dl75s\" (UniqueName: \"kubernetes.io/projected/d6e90da6-e926-4a6a-8a61-a418b5cd15d7-kube-api-access-dl75s\") pod \"d6e90da6-e926-4a6a-8a61-a418b5cd15d7\" (UID: \"d6e90da6-e926-4a6a-8a61-a418b5cd15d7\") " Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.085608 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6e90da6-e926-4a6a-8a61-a418b5cd15d7-catalog-content\") pod \"d6e90da6-e926-4a6a-8a61-a418b5cd15d7\" (UID: \"d6e90da6-e926-4a6a-8a61-a418b5cd15d7\") " Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.085965 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6e90da6-e926-4a6a-8a61-a418b5cd15d7-utilities\") pod \"d6e90da6-e926-4a6a-8a61-a418b5cd15d7\" (UID: \"d6e90da6-e926-4a6a-8a61-a418b5cd15d7\") " Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.086409 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6e90da6-e926-4a6a-8a61-a418b5cd15d7-utilities" (OuterVolumeSpecName: "utilities") pod "d6e90da6-e926-4a6a-8a61-a418b5cd15d7" (UID: "d6e90da6-e926-4a6a-8a61-a418b5cd15d7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.086995 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6e90da6-e926-4a6a-8a61-a418b5cd15d7-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.091812 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6e90da6-e926-4a6a-8a61-a418b5cd15d7-kube-api-access-dl75s" (OuterVolumeSpecName: "kube-api-access-dl75s") pod "d6e90da6-e926-4a6a-8a61-a418b5cd15d7" (UID: "d6e90da6-e926-4a6a-8a61-a418b5cd15d7"). InnerVolumeSpecName "kube-api-access-dl75s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.135555 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6e90da6-e926-4a6a-8a61-a418b5cd15d7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d6e90da6-e926-4a6a-8a61-a418b5cd15d7" (UID: "d6e90da6-e926-4a6a-8a61-a418b5cd15d7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.188726 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dl75s\" (UniqueName: \"kubernetes.io/projected/d6e90da6-e926-4a6a-8a61-a418b5cd15d7-kube-api-access-dl75s\") on node \"crc\" DevicePath \"\"" Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.188769 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6e90da6-e926-4a6a-8a61-a418b5cd15d7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.462800 4934 generic.go:334] "Generic (PLEG): container finished" podID="d6e90da6-e926-4a6a-8a61-a418b5cd15d7" containerID="450bb0fa53d0f90dfa7edb99cac4e09c3c4a60ceff0584d3d3e446aeea40e077" exitCode=0 Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.462865 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2sc7g" event={"ID":"d6e90da6-e926-4a6a-8a61-a418b5cd15d7","Type":"ContainerDied","Data":"450bb0fa53d0f90dfa7edb99cac4e09c3c4a60ceff0584d3d3e446aeea40e077"} Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.462902 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2sc7g" event={"ID":"d6e90da6-e926-4a6a-8a61-a418b5cd15d7","Type":"ContainerDied","Data":"1dd0eb9520faf01e7de43da8093771ccb32e6c02b8b8c4753ee308e4df4f94c5"} Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.462945 4934 scope.go:117] "RemoveContainer" containerID="450bb0fa53d0f90dfa7edb99cac4e09c3c4a60ceff0584d3d3e446aeea40e077" Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.463217 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2sc7g" Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.509956 4934 scope.go:117] "RemoveContainer" containerID="bc4c1de8c2f2f32872a107e4d05a4814cd0766b49d405a1ac8e949473712f430" Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.522759 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2sc7g"] Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.533876 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2sc7g"] Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.536455 4934 scope.go:117] "RemoveContainer" containerID="5d2fcb8d4c013561b4b3cfe12e0a97dc8f28694d8fdf5f9ea61a2eccd224ffe8" Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.596789 4934 scope.go:117] "RemoveContainer" containerID="450bb0fa53d0f90dfa7edb99cac4e09c3c4a60ceff0584d3d3e446aeea40e077" Dec 27 09:17:59 crc kubenswrapper[4934]: E1227 09:17:59.597731 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"450bb0fa53d0f90dfa7edb99cac4e09c3c4a60ceff0584d3d3e446aeea40e077\": container with ID starting with 450bb0fa53d0f90dfa7edb99cac4e09c3c4a60ceff0584d3d3e446aeea40e077 not found: ID does not exist" containerID="450bb0fa53d0f90dfa7edb99cac4e09c3c4a60ceff0584d3d3e446aeea40e077" Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.597800 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"450bb0fa53d0f90dfa7edb99cac4e09c3c4a60ceff0584d3d3e446aeea40e077"} err="failed to get container status \"450bb0fa53d0f90dfa7edb99cac4e09c3c4a60ceff0584d3d3e446aeea40e077\": rpc error: code = NotFound desc = could not find container \"450bb0fa53d0f90dfa7edb99cac4e09c3c4a60ceff0584d3d3e446aeea40e077\": container with ID starting with 450bb0fa53d0f90dfa7edb99cac4e09c3c4a60ceff0584d3d3e446aeea40e077 not found: ID does not exist" Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.597850 4934 scope.go:117] "RemoveContainer" containerID="bc4c1de8c2f2f32872a107e4d05a4814cd0766b49d405a1ac8e949473712f430" Dec 27 09:17:59 crc kubenswrapper[4934]: E1227 09:17:59.598731 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc4c1de8c2f2f32872a107e4d05a4814cd0766b49d405a1ac8e949473712f430\": container with ID starting with bc4c1de8c2f2f32872a107e4d05a4814cd0766b49d405a1ac8e949473712f430 not found: ID does not exist" containerID="bc4c1de8c2f2f32872a107e4d05a4814cd0766b49d405a1ac8e949473712f430" Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.598820 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc4c1de8c2f2f32872a107e4d05a4814cd0766b49d405a1ac8e949473712f430"} err="failed to get container status \"bc4c1de8c2f2f32872a107e4d05a4814cd0766b49d405a1ac8e949473712f430\": rpc error: code = NotFound desc = could not find container \"bc4c1de8c2f2f32872a107e4d05a4814cd0766b49d405a1ac8e949473712f430\": container with ID starting with bc4c1de8c2f2f32872a107e4d05a4814cd0766b49d405a1ac8e949473712f430 not found: ID does not exist" Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.598865 4934 scope.go:117] "RemoveContainer" containerID="5d2fcb8d4c013561b4b3cfe12e0a97dc8f28694d8fdf5f9ea61a2eccd224ffe8" Dec 27 09:17:59 crc kubenswrapper[4934]: E1227 09:17:59.599334 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d2fcb8d4c013561b4b3cfe12e0a97dc8f28694d8fdf5f9ea61a2eccd224ffe8\": container with ID starting with 5d2fcb8d4c013561b4b3cfe12e0a97dc8f28694d8fdf5f9ea61a2eccd224ffe8 not found: ID does not exist" containerID="5d2fcb8d4c013561b4b3cfe12e0a97dc8f28694d8fdf5f9ea61a2eccd224ffe8" Dec 27 09:17:59 crc kubenswrapper[4934]: I1227 09:17:59.599361 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d2fcb8d4c013561b4b3cfe12e0a97dc8f28694d8fdf5f9ea61a2eccd224ffe8"} err="failed to get container status \"5d2fcb8d4c013561b4b3cfe12e0a97dc8f28694d8fdf5f9ea61a2eccd224ffe8\": rpc error: code = NotFound desc = could not find container \"5d2fcb8d4c013561b4b3cfe12e0a97dc8f28694d8fdf5f9ea61a2eccd224ffe8\": container with ID starting with 5d2fcb8d4c013561b4b3cfe12e0a97dc8f28694d8fdf5f9ea61a2eccd224ffe8 not found: ID does not exist" Dec 27 09:18:01 crc kubenswrapper[4934]: I1227 09:18:01.495938 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6e90da6-e926-4a6a-8a61-a418b5cd15d7" path="/var/lib/kubelet/pods/d6e90da6-e926-4a6a-8a61-a418b5cd15d7/volumes" Dec 27 09:18:01 crc kubenswrapper[4934]: I1227 09:18:01.982790 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hzlkj" Dec 27 09:18:01 crc kubenswrapper[4934]: I1227 09:18:01.982878 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hzlkj" Dec 27 09:18:02 crc kubenswrapper[4934]: I1227 09:18:02.049157 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hzlkj" Dec 27 09:18:02 crc kubenswrapper[4934]: I1227 09:18:02.570929 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hzlkj" Dec 27 09:18:03 crc kubenswrapper[4934]: I1227 09:18:03.191878 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hzlkj"] Dec 27 09:18:04 crc kubenswrapper[4934]: I1227 09:18:04.534281 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hzlkj" podUID="f972030f-cb58-420b-927e-80ea4ccb198b" containerName="registry-server" containerID="cri-o://7ef2085cf460dc4930f4658c8290f48e38cc30e41bd78760b12ecee5605fbb3b" gracePeriod=2 Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.373884 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hzlkj" Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.442367 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f972030f-cb58-420b-927e-80ea4ccb198b-utilities\") pod \"f972030f-cb58-420b-927e-80ea4ccb198b\" (UID: \"f972030f-cb58-420b-927e-80ea4ccb198b\") " Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.442421 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xv85q\" (UniqueName: \"kubernetes.io/projected/f972030f-cb58-420b-927e-80ea4ccb198b-kube-api-access-xv85q\") pod \"f972030f-cb58-420b-927e-80ea4ccb198b\" (UID: \"f972030f-cb58-420b-927e-80ea4ccb198b\") " Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.442450 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f972030f-cb58-420b-927e-80ea4ccb198b-catalog-content\") pod \"f972030f-cb58-420b-927e-80ea4ccb198b\" (UID: \"f972030f-cb58-420b-927e-80ea4ccb198b\") " Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.443099 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f972030f-cb58-420b-927e-80ea4ccb198b-utilities" (OuterVolumeSpecName: "utilities") pod "f972030f-cb58-420b-927e-80ea4ccb198b" (UID: "f972030f-cb58-420b-927e-80ea4ccb198b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.443415 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f972030f-cb58-420b-927e-80ea4ccb198b-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.450251 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f972030f-cb58-420b-927e-80ea4ccb198b-kube-api-access-xv85q" (OuterVolumeSpecName: "kube-api-access-xv85q") pod "f972030f-cb58-420b-927e-80ea4ccb198b" (UID: "f972030f-cb58-420b-927e-80ea4ccb198b"). InnerVolumeSpecName "kube-api-access-xv85q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.500575 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f972030f-cb58-420b-927e-80ea4ccb198b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f972030f-cb58-420b-927e-80ea4ccb198b" (UID: "f972030f-cb58-420b-927e-80ea4ccb198b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.545840 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xv85q\" (UniqueName: \"kubernetes.io/projected/f972030f-cb58-420b-927e-80ea4ccb198b-kube-api-access-xv85q\") on node \"crc\" DevicePath \"\"" Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.545877 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f972030f-cb58-420b-927e-80ea4ccb198b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.550588 4934 generic.go:334] "Generic (PLEG): container finished" podID="f972030f-cb58-420b-927e-80ea4ccb198b" containerID="7ef2085cf460dc4930f4658c8290f48e38cc30e41bd78760b12ecee5605fbb3b" exitCode=0 Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.550641 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzlkj" event={"ID":"f972030f-cb58-420b-927e-80ea4ccb198b","Type":"ContainerDied","Data":"7ef2085cf460dc4930f4658c8290f48e38cc30e41bd78760b12ecee5605fbb3b"} Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.550662 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hzlkj" Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.550681 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzlkj" event={"ID":"f972030f-cb58-420b-927e-80ea4ccb198b","Type":"ContainerDied","Data":"0fdc9e4e804491916b3f8fba768e9c13db530c2cd7dc7d3aca95dcb8624e7e10"} Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.550705 4934 scope.go:117] "RemoveContainer" containerID="7ef2085cf460dc4930f4658c8290f48e38cc30e41bd78760b12ecee5605fbb3b" Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.578290 4934 scope.go:117] "RemoveContainer" containerID="77d33799b02558dd1e41317bbf461aee0ad5d7d0780c0e3ec14583fdb19fe23b" Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.595180 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hzlkj"] Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.608306 4934 scope.go:117] "RemoveContainer" containerID="72b880a5187c780e9810bba436455bfd3701aabe5aa7eb3b00ce71ea1711f3cb" Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.615420 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hzlkj"] Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.664533 4934 scope.go:117] "RemoveContainer" containerID="7ef2085cf460dc4930f4658c8290f48e38cc30e41bd78760b12ecee5605fbb3b" Dec 27 09:18:05 crc kubenswrapper[4934]: E1227 09:18:05.664976 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ef2085cf460dc4930f4658c8290f48e38cc30e41bd78760b12ecee5605fbb3b\": container with ID starting with 7ef2085cf460dc4930f4658c8290f48e38cc30e41bd78760b12ecee5605fbb3b not found: ID does not exist" containerID="7ef2085cf460dc4930f4658c8290f48e38cc30e41bd78760b12ecee5605fbb3b" Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.665012 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ef2085cf460dc4930f4658c8290f48e38cc30e41bd78760b12ecee5605fbb3b"} err="failed to get container status \"7ef2085cf460dc4930f4658c8290f48e38cc30e41bd78760b12ecee5605fbb3b\": rpc error: code = NotFound desc = could not find container \"7ef2085cf460dc4930f4658c8290f48e38cc30e41bd78760b12ecee5605fbb3b\": container with ID starting with 7ef2085cf460dc4930f4658c8290f48e38cc30e41bd78760b12ecee5605fbb3b not found: ID does not exist" Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.665037 4934 scope.go:117] "RemoveContainer" containerID="77d33799b02558dd1e41317bbf461aee0ad5d7d0780c0e3ec14583fdb19fe23b" Dec 27 09:18:05 crc kubenswrapper[4934]: E1227 09:18:05.665471 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77d33799b02558dd1e41317bbf461aee0ad5d7d0780c0e3ec14583fdb19fe23b\": container with ID starting with 77d33799b02558dd1e41317bbf461aee0ad5d7d0780c0e3ec14583fdb19fe23b not found: ID does not exist" containerID="77d33799b02558dd1e41317bbf461aee0ad5d7d0780c0e3ec14583fdb19fe23b" Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.665575 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77d33799b02558dd1e41317bbf461aee0ad5d7d0780c0e3ec14583fdb19fe23b"} err="failed to get container status \"77d33799b02558dd1e41317bbf461aee0ad5d7d0780c0e3ec14583fdb19fe23b\": rpc error: code = NotFound desc = could not find container \"77d33799b02558dd1e41317bbf461aee0ad5d7d0780c0e3ec14583fdb19fe23b\": container with ID starting with 77d33799b02558dd1e41317bbf461aee0ad5d7d0780c0e3ec14583fdb19fe23b not found: ID does not exist" Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.665660 4934 scope.go:117] "RemoveContainer" containerID="72b880a5187c780e9810bba436455bfd3701aabe5aa7eb3b00ce71ea1711f3cb" Dec 27 09:18:05 crc kubenswrapper[4934]: E1227 09:18:05.666203 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72b880a5187c780e9810bba436455bfd3701aabe5aa7eb3b00ce71ea1711f3cb\": container with ID starting with 72b880a5187c780e9810bba436455bfd3701aabe5aa7eb3b00ce71ea1711f3cb not found: ID does not exist" containerID="72b880a5187c780e9810bba436455bfd3701aabe5aa7eb3b00ce71ea1711f3cb" Dec 27 09:18:05 crc kubenswrapper[4934]: I1227 09:18:05.666419 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72b880a5187c780e9810bba436455bfd3701aabe5aa7eb3b00ce71ea1711f3cb"} err="failed to get container status \"72b880a5187c780e9810bba436455bfd3701aabe5aa7eb3b00ce71ea1711f3cb\": rpc error: code = NotFound desc = could not find container \"72b880a5187c780e9810bba436455bfd3701aabe5aa7eb3b00ce71ea1711f3cb\": container with ID starting with 72b880a5187c780e9810bba436455bfd3701aabe5aa7eb3b00ce71ea1711f3cb not found: ID does not exist" Dec 27 09:18:07 crc kubenswrapper[4934]: I1227 09:18:07.487546 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f972030f-cb58-420b-927e-80ea4ccb198b" path="/var/lib/kubelet/pods/f972030f-cb58-420b-927e-80ea4ccb198b/volumes" Dec 27 09:18:10 crc kubenswrapper[4934]: I1227 09:18:10.468748 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:18:10 crc kubenswrapper[4934]: E1227 09:18:10.469643 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:18:18 crc kubenswrapper[4934]: I1227 09:18:18.125573 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jqg5s"] Dec 27 09:18:18 crc kubenswrapper[4934]: E1227 09:18:18.127946 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e90da6-e926-4a6a-8a61-a418b5cd15d7" containerName="extract-content" Dec 27 09:18:18 crc kubenswrapper[4934]: I1227 09:18:18.128132 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e90da6-e926-4a6a-8a61-a418b5cd15d7" containerName="extract-content" Dec 27 09:18:18 crc kubenswrapper[4934]: E1227 09:18:18.128310 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f972030f-cb58-420b-927e-80ea4ccb198b" containerName="registry-server" Dec 27 09:18:18 crc kubenswrapper[4934]: I1227 09:18:18.128455 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f972030f-cb58-420b-927e-80ea4ccb198b" containerName="registry-server" Dec 27 09:18:18 crc kubenswrapper[4934]: E1227 09:18:18.128589 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f972030f-cb58-420b-927e-80ea4ccb198b" containerName="extract-content" Dec 27 09:18:18 crc kubenswrapper[4934]: I1227 09:18:18.128680 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f972030f-cb58-420b-927e-80ea4ccb198b" containerName="extract-content" Dec 27 09:18:18 crc kubenswrapper[4934]: E1227 09:18:18.128783 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e90da6-e926-4a6a-8a61-a418b5cd15d7" containerName="registry-server" Dec 27 09:18:18 crc kubenswrapper[4934]: I1227 09:18:18.128875 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e90da6-e926-4a6a-8a61-a418b5cd15d7" containerName="registry-server" Dec 27 09:18:18 crc kubenswrapper[4934]: E1227 09:18:18.129007 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e90da6-e926-4a6a-8a61-a418b5cd15d7" containerName="extract-utilities" Dec 27 09:18:18 crc kubenswrapper[4934]: I1227 09:18:18.129114 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e90da6-e926-4a6a-8a61-a418b5cd15d7" containerName="extract-utilities" Dec 27 09:18:18 crc kubenswrapper[4934]: E1227 09:18:18.129290 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f972030f-cb58-420b-927e-80ea4ccb198b" containerName="extract-utilities" Dec 27 09:18:18 crc kubenswrapper[4934]: I1227 09:18:18.129378 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="f972030f-cb58-420b-927e-80ea4ccb198b" containerName="extract-utilities" Dec 27 09:18:18 crc kubenswrapper[4934]: I1227 09:18:18.129763 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="f972030f-cb58-420b-927e-80ea4ccb198b" containerName="registry-server" Dec 27 09:18:18 crc kubenswrapper[4934]: I1227 09:18:18.129897 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6e90da6-e926-4a6a-8a61-a418b5cd15d7" containerName="registry-server" Dec 27 09:18:18 crc kubenswrapper[4934]: I1227 09:18:18.132551 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jqg5s" Dec 27 09:18:18 crc kubenswrapper[4934]: I1227 09:18:18.139511 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jqg5s"] Dec 27 09:18:18 crc kubenswrapper[4934]: I1227 09:18:18.205417 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/409de99a-6cd8-4574-8727-d59a71321d8d-catalog-content\") pod \"redhat-operators-jqg5s\" (UID: \"409de99a-6cd8-4574-8727-d59a71321d8d\") " pod="openshift-marketplace/redhat-operators-jqg5s" Dec 27 09:18:18 crc kubenswrapper[4934]: I1227 09:18:18.205559 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xtxb\" (UniqueName: \"kubernetes.io/projected/409de99a-6cd8-4574-8727-d59a71321d8d-kube-api-access-7xtxb\") pod \"redhat-operators-jqg5s\" (UID: \"409de99a-6cd8-4574-8727-d59a71321d8d\") " pod="openshift-marketplace/redhat-operators-jqg5s" Dec 27 09:18:18 crc kubenswrapper[4934]: I1227 09:18:18.205631 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/409de99a-6cd8-4574-8727-d59a71321d8d-utilities\") pod \"redhat-operators-jqg5s\" (UID: \"409de99a-6cd8-4574-8727-d59a71321d8d\") " pod="openshift-marketplace/redhat-operators-jqg5s" Dec 27 09:18:18 crc kubenswrapper[4934]: I1227 09:18:18.308071 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/409de99a-6cd8-4574-8727-d59a71321d8d-catalog-content\") pod \"redhat-operators-jqg5s\" (UID: \"409de99a-6cd8-4574-8727-d59a71321d8d\") " pod="openshift-marketplace/redhat-operators-jqg5s" Dec 27 09:18:18 crc kubenswrapper[4934]: I1227 09:18:18.308282 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xtxb\" (UniqueName: \"kubernetes.io/projected/409de99a-6cd8-4574-8727-d59a71321d8d-kube-api-access-7xtxb\") pod \"redhat-operators-jqg5s\" (UID: \"409de99a-6cd8-4574-8727-d59a71321d8d\") " pod="openshift-marketplace/redhat-operators-jqg5s" Dec 27 09:18:18 crc kubenswrapper[4934]: I1227 09:18:18.308376 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/409de99a-6cd8-4574-8727-d59a71321d8d-utilities\") pod \"redhat-operators-jqg5s\" (UID: \"409de99a-6cd8-4574-8727-d59a71321d8d\") " pod="openshift-marketplace/redhat-operators-jqg5s" Dec 27 09:18:18 crc kubenswrapper[4934]: I1227 09:18:18.308623 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/409de99a-6cd8-4574-8727-d59a71321d8d-catalog-content\") pod \"redhat-operators-jqg5s\" (UID: \"409de99a-6cd8-4574-8727-d59a71321d8d\") " pod="openshift-marketplace/redhat-operators-jqg5s" Dec 27 09:18:18 crc kubenswrapper[4934]: I1227 09:18:18.309256 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/409de99a-6cd8-4574-8727-d59a71321d8d-utilities\") pod \"redhat-operators-jqg5s\" (UID: \"409de99a-6cd8-4574-8727-d59a71321d8d\") " pod="openshift-marketplace/redhat-operators-jqg5s" Dec 27 09:18:18 crc kubenswrapper[4934]: I1227 09:18:18.546349 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xtxb\" (UniqueName: \"kubernetes.io/projected/409de99a-6cd8-4574-8727-d59a71321d8d-kube-api-access-7xtxb\") pod \"redhat-operators-jqg5s\" (UID: \"409de99a-6cd8-4574-8727-d59a71321d8d\") " pod="openshift-marketplace/redhat-operators-jqg5s" Dec 27 09:18:18 crc kubenswrapper[4934]: I1227 09:18:18.769171 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jqg5s" Dec 27 09:18:19 crc kubenswrapper[4934]: I1227 09:18:19.288469 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jqg5s"] Dec 27 09:18:19 crc kubenswrapper[4934]: I1227 09:18:19.732128 4934 generic.go:334] "Generic (PLEG): container finished" podID="409de99a-6cd8-4574-8727-d59a71321d8d" containerID="55dcd3da63e6bba81b9e89f34fbebadf34ab438239d03d13e61132eaf2f89ecc" exitCode=0 Dec 27 09:18:19 crc kubenswrapper[4934]: I1227 09:18:19.732204 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jqg5s" event={"ID":"409de99a-6cd8-4574-8727-d59a71321d8d","Type":"ContainerDied","Data":"55dcd3da63e6bba81b9e89f34fbebadf34ab438239d03d13e61132eaf2f89ecc"} Dec 27 09:18:19 crc kubenswrapper[4934]: I1227 09:18:19.732473 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jqg5s" event={"ID":"409de99a-6cd8-4574-8727-d59a71321d8d","Type":"ContainerStarted","Data":"f574bec91d60d052207b4cef6db920983b94d5534d7d3d50b3cd41540c07bb99"} Dec 27 09:18:21 crc kubenswrapper[4934]: I1227 09:18:21.764429 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jqg5s" event={"ID":"409de99a-6cd8-4574-8727-d59a71321d8d","Type":"ContainerStarted","Data":"bac3b1750a49cc4d2e80862c26063bc87029bd565f2ba33e23477eb00d27e221"} Dec 27 09:18:21 crc kubenswrapper[4934]: I1227 09:18:21.905132 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-m4qwr"] Dec 27 09:18:21 crc kubenswrapper[4934]: I1227 09:18:21.909925 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m4qwr" Dec 27 09:18:21 crc kubenswrapper[4934]: I1227 09:18:21.934385 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m4qwr"] Dec 27 09:18:22 crc kubenswrapper[4934]: I1227 09:18:22.009469 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae6cb076-b837-4671-83ba-43d479b94443-catalog-content\") pod \"redhat-marketplace-m4qwr\" (UID: \"ae6cb076-b837-4671-83ba-43d479b94443\") " pod="openshift-marketplace/redhat-marketplace-m4qwr" Dec 27 09:18:22 crc kubenswrapper[4934]: I1227 09:18:22.009924 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae6cb076-b837-4671-83ba-43d479b94443-utilities\") pod \"redhat-marketplace-m4qwr\" (UID: \"ae6cb076-b837-4671-83ba-43d479b94443\") " pod="openshift-marketplace/redhat-marketplace-m4qwr" Dec 27 09:18:22 crc kubenswrapper[4934]: I1227 09:18:22.009952 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rj4vt\" (UniqueName: \"kubernetes.io/projected/ae6cb076-b837-4671-83ba-43d479b94443-kube-api-access-rj4vt\") pod \"redhat-marketplace-m4qwr\" (UID: \"ae6cb076-b837-4671-83ba-43d479b94443\") " pod="openshift-marketplace/redhat-marketplace-m4qwr" Dec 27 09:18:22 crc kubenswrapper[4934]: I1227 09:18:22.112502 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae6cb076-b837-4671-83ba-43d479b94443-catalog-content\") pod \"redhat-marketplace-m4qwr\" (UID: \"ae6cb076-b837-4671-83ba-43d479b94443\") " pod="openshift-marketplace/redhat-marketplace-m4qwr" Dec 27 09:18:22 crc kubenswrapper[4934]: I1227 09:18:22.112610 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae6cb076-b837-4671-83ba-43d479b94443-utilities\") pod \"redhat-marketplace-m4qwr\" (UID: \"ae6cb076-b837-4671-83ba-43d479b94443\") " pod="openshift-marketplace/redhat-marketplace-m4qwr" Dec 27 09:18:22 crc kubenswrapper[4934]: I1227 09:18:22.112629 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rj4vt\" (UniqueName: \"kubernetes.io/projected/ae6cb076-b837-4671-83ba-43d479b94443-kube-api-access-rj4vt\") pod \"redhat-marketplace-m4qwr\" (UID: \"ae6cb076-b837-4671-83ba-43d479b94443\") " pod="openshift-marketplace/redhat-marketplace-m4qwr" Dec 27 09:18:22 crc kubenswrapper[4934]: I1227 09:18:22.113067 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae6cb076-b837-4671-83ba-43d479b94443-catalog-content\") pod \"redhat-marketplace-m4qwr\" (UID: \"ae6cb076-b837-4671-83ba-43d479b94443\") " pod="openshift-marketplace/redhat-marketplace-m4qwr" Dec 27 09:18:22 crc kubenswrapper[4934]: I1227 09:18:22.113074 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae6cb076-b837-4671-83ba-43d479b94443-utilities\") pod \"redhat-marketplace-m4qwr\" (UID: \"ae6cb076-b837-4671-83ba-43d479b94443\") " pod="openshift-marketplace/redhat-marketplace-m4qwr" Dec 27 09:18:22 crc kubenswrapper[4934]: I1227 09:18:22.132066 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rj4vt\" (UniqueName: \"kubernetes.io/projected/ae6cb076-b837-4671-83ba-43d479b94443-kube-api-access-rj4vt\") pod \"redhat-marketplace-m4qwr\" (UID: \"ae6cb076-b837-4671-83ba-43d479b94443\") " pod="openshift-marketplace/redhat-marketplace-m4qwr" Dec 27 09:18:22 crc kubenswrapper[4934]: I1227 09:18:22.235278 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m4qwr" Dec 27 09:18:22 crc kubenswrapper[4934]: I1227 09:18:22.467357 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:18:22 crc kubenswrapper[4934]: E1227 09:18:22.467836 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:18:22 crc kubenswrapper[4934]: I1227 09:18:22.784909 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-m4qwr"] Dec 27 09:18:22 crc kubenswrapper[4934]: W1227 09:18:22.880026 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae6cb076_b837_4671_83ba_43d479b94443.slice/crio-e6e42f691e524d38f65126945d52cd3939d4f65c49391cd561be9444cb025b9d WatchSource:0}: Error finding container e6e42f691e524d38f65126945d52cd3939d4f65c49391cd561be9444cb025b9d: Status 404 returned error can't find the container with id e6e42f691e524d38f65126945d52cd3939d4f65c49391cd561be9444cb025b9d Dec 27 09:18:23 crc kubenswrapper[4934]: I1227 09:18:23.809580 4934 generic.go:334] "Generic (PLEG): container finished" podID="ae6cb076-b837-4671-83ba-43d479b94443" containerID="12764a36c08b67337808b28e847cc266ca1713e84de57f838504d5c559993b85" exitCode=0 Dec 27 09:18:23 crc kubenswrapper[4934]: I1227 09:18:23.809845 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m4qwr" event={"ID":"ae6cb076-b837-4671-83ba-43d479b94443","Type":"ContainerDied","Data":"12764a36c08b67337808b28e847cc266ca1713e84de57f838504d5c559993b85"} Dec 27 09:18:23 crc kubenswrapper[4934]: I1227 09:18:23.809874 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m4qwr" event={"ID":"ae6cb076-b837-4671-83ba-43d479b94443","Type":"ContainerStarted","Data":"e6e42f691e524d38f65126945d52cd3939d4f65c49391cd561be9444cb025b9d"} Dec 27 09:18:24 crc kubenswrapper[4934]: I1227 09:18:24.826396 4934 generic.go:334] "Generic (PLEG): container finished" podID="409de99a-6cd8-4574-8727-d59a71321d8d" containerID="bac3b1750a49cc4d2e80862c26063bc87029bd565f2ba33e23477eb00d27e221" exitCode=0 Dec 27 09:18:24 crc kubenswrapper[4934]: I1227 09:18:24.826491 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jqg5s" event={"ID":"409de99a-6cd8-4574-8727-d59a71321d8d","Type":"ContainerDied","Data":"bac3b1750a49cc4d2e80862c26063bc87029bd565f2ba33e23477eb00d27e221"} Dec 27 09:18:25 crc kubenswrapper[4934]: I1227 09:18:25.839572 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m4qwr" event={"ID":"ae6cb076-b837-4671-83ba-43d479b94443","Type":"ContainerStarted","Data":"b0482879005ff56387cb38df22c6a3856d9b92fc44414f4ba7c4df70aa7d5a2c"} Dec 27 09:18:26 crc kubenswrapper[4934]: I1227 09:18:26.855773 4934 generic.go:334] "Generic (PLEG): container finished" podID="ae6cb076-b837-4671-83ba-43d479b94443" containerID="b0482879005ff56387cb38df22c6a3856d9b92fc44414f4ba7c4df70aa7d5a2c" exitCode=0 Dec 27 09:18:26 crc kubenswrapper[4934]: I1227 09:18:26.856064 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m4qwr" event={"ID":"ae6cb076-b837-4671-83ba-43d479b94443","Type":"ContainerDied","Data":"b0482879005ff56387cb38df22c6a3856d9b92fc44414f4ba7c4df70aa7d5a2c"} Dec 27 09:18:26 crc kubenswrapper[4934]: I1227 09:18:26.860047 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jqg5s" event={"ID":"409de99a-6cd8-4574-8727-d59a71321d8d","Type":"ContainerStarted","Data":"0db1a02b1ca44a9e50336ee70f02ff1246ea0200ffddfd7859d1b6a08b86ca36"} Dec 27 09:18:26 crc kubenswrapper[4934]: I1227 09:18:26.909124 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jqg5s" podStartSLOduration=2.088923437 podStartE2EDuration="8.909104552s" podCreationTimestamp="2025-12-27 09:18:18 +0000 UTC" firstStartedPulling="2025-12-27 09:18:19.736173488 +0000 UTC m=+5760.556614092" lastFinishedPulling="2025-12-27 09:18:26.556354583 +0000 UTC m=+5767.376795207" observedRunningTime="2025-12-27 09:18:26.896459159 +0000 UTC m=+5767.716899773" watchObservedRunningTime="2025-12-27 09:18:26.909104552 +0000 UTC m=+5767.729545146" Dec 27 09:18:27 crc kubenswrapper[4934]: I1227 09:18:27.874048 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m4qwr" event={"ID":"ae6cb076-b837-4671-83ba-43d479b94443","Type":"ContainerStarted","Data":"9fb9a9a105f82b1c3ef2cdc89348c9929df9be11c73f572ecd9fa03130a17376"} Dec 27 09:18:27 crc kubenswrapper[4934]: I1227 09:18:27.904833 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-m4qwr" podStartSLOduration=3.210117262 podStartE2EDuration="6.904810663s" podCreationTimestamp="2025-12-27 09:18:21 +0000 UTC" firstStartedPulling="2025-12-27 09:18:23.812291877 +0000 UTC m=+5764.632732471" lastFinishedPulling="2025-12-27 09:18:27.506985278 +0000 UTC m=+5768.327425872" observedRunningTime="2025-12-27 09:18:27.899368438 +0000 UTC m=+5768.719809062" watchObservedRunningTime="2025-12-27 09:18:27.904810663 +0000 UTC m=+5768.725251257" Dec 27 09:18:28 crc kubenswrapper[4934]: I1227 09:18:28.770116 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jqg5s" Dec 27 09:18:28 crc kubenswrapper[4934]: I1227 09:18:28.770439 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jqg5s" Dec 27 09:18:29 crc kubenswrapper[4934]: I1227 09:18:29.831341 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jqg5s" podUID="409de99a-6cd8-4574-8727-d59a71321d8d" containerName="registry-server" probeResult="failure" output=< Dec 27 09:18:29 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 09:18:29 crc kubenswrapper[4934]: > Dec 27 09:18:32 crc kubenswrapper[4934]: I1227 09:18:32.235691 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-m4qwr" Dec 27 09:18:32 crc kubenswrapper[4934]: I1227 09:18:32.236931 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-m4qwr" Dec 27 09:18:32 crc kubenswrapper[4934]: I1227 09:18:32.300823 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-m4qwr" Dec 27 09:18:33 crc kubenswrapper[4934]: I1227 09:18:33.015118 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-m4qwr" Dec 27 09:18:33 crc kubenswrapper[4934]: I1227 09:18:33.072744 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m4qwr"] Dec 27 09:18:33 crc kubenswrapper[4934]: I1227 09:18:33.467152 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:18:33 crc kubenswrapper[4934]: E1227 09:18:33.467411 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:18:34 crc kubenswrapper[4934]: I1227 09:18:34.960215 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-m4qwr" podUID="ae6cb076-b837-4671-83ba-43d479b94443" containerName="registry-server" containerID="cri-o://9fb9a9a105f82b1c3ef2cdc89348c9929df9be11c73f572ecd9fa03130a17376" gracePeriod=2 Dec 27 09:18:35 crc kubenswrapper[4934]: I1227 09:18:35.973421 4934 generic.go:334] "Generic (PLEG): container finished" podID="ae6cb076-b837-4671-83ba-43d479b94443" containerID="9fb9a9a105f82b1c3ef2cdc89348c9929df9be11c73f572ecd9fa03130a17376" exitCode=0 Dec 27 09:18:35 crc kubenswrapper[4934]: I1227 09:18:35.973500 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m4qwr" event={"ID":"ae6cb076-b837-4671-83ba-43d479b94443","Type":"ContainerDied","Data":"9fb9a9a105f82b1c3ef2cdc89348c9929df9be11c73f572ecd9fa03130a17376"} Dec 27 09:18:35 crc kubenswrapper[4934]: I1227 09:18:35.973683 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-m4qwr" event={"ID":"ae6cb076-b837-4671-83ba-43d479b94443","Type":"ContainerDied","Data":"e6e42f691e524d38f65126945d52cd3939d4f65c49391cd561be9444cb025b9d"} Dec 27 09:18:35 crc kubenswrapper[4934]: I1227 09:18:35.973695 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6e42f691e524d38f65126945d52cd3939d4f65c49391cd561be9444cb025b9d" Dec 27 09:18:36 crc kubenswrapper[4934]: I1227 09:18:36.801945 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m4qwr" Dec 27 09:18:36 crc kubenswrapper[4934]: I1227 09:18:36.883247 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae6cb076-b837-4671-83ba-43d479b94443-utilities\") pod \"ae6cb076-b837-4671-83ba-43d479b94443\" (UID: \"ae6cb076-b837-4671-83ba-43d479b94443\") " Dec 27 09:18:36 crc kubenswrapper[4934]: I1227 09:18:36.883301 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rj4vt\" (UniqueName: \"kubernetes.io/projected/ae6cb076-b837-4671-83ba-43d479b94443-kube-api-access-rj4vt\") pod \"ae6cb076-b837-4671-83ba-43d479b94443\" (UID: \"ae6cb076-b837-4671-83ba-43d479b94443\") " Dec 27 09:18:36 crc kubenswrapper[4934]: I1227 09:18:36.883582 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae6cb076-b837-4671-83ba-43d479b94443-catalog-content\") pod \"ae6cb076-b837-4671-83ba-43d479b94443\" (UID: \"ae6cb076-b837-4671-83ba-43d479b94443\") " Dec 27 09:18:36 crc kubenswrapper[4934]: I1227 09:18:36.884097 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae6cb076-b837-4671-83ba-43d479b94443-utilities" (OuterVolumeSpecName: "utilities") pod "ae6cb076-b837-4671-83ba-43d479b94443" (UID: "ae6cb076-b837-4671-83ba-43d479b94443"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:18:36 crc kubenswrapper[4934]: I1227 09:18:36.884445 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae6cb076-b837-4671-83ba-43d479b94443-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 09:18:36 crc kubenswrapper[4934]: I1227 09:18:36.900799 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae6cb076-b837-4671-83ba-43d479b94443-kube-api-access-rj4vt" (OuterVolumeSpecName: "kube-api-access-rj4vt") pod "ae6cb076-b837-4671-83ba-43d479b94443" (UID: "ae6cb076-b837-4671-83ba-43d479b94443"). InnerVolumeSpecName "kube-api-access-rj4vt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:18:36 crc kubenswrapper[4934]: I1227 09:18:36.903535 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae6cb076-b837-4671-83ba-43d479b94443-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae6cb076-b837-4671-83ba-43d479b94443" (UID: "ae6cb076-b837-4671-83ba-43d479b94443"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:18:36 crc kubenswrapper[4934]: I1227 09:18:36.984234 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-m4qwr" Dec 27 09:18:36 crc kubenswrapper[4934]: I1227 09:18:36.986574 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae6cb076-b837-4671-83ba-43d479b94443-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 09:18:36 crc kubenswrapper[4934]: I1227 09:18:36.986613 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rj4vt\" (UniqueName: \"kubernetes.io/projected/ae6cb076-b837-4671-83ba-43d479b94443-kube-api-access-rj4vt\") on node \"crc\" DevicePath \"\"" Dec 27 09:18:37 crc kubenswrapper[4934]: I1227 09:18:37.019961 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-m4qwr"] Dec 27 09:18:37 crc kubenswrapper[4934]: I1227 09:18:37.029937 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-m4qwr"] Dec 27 09:18:37 crc kubenswrapper[4934]: I1227 09:18:37.482273 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae6cb076-b837-4671-83ba-43d479b94443" path="/var/lib/kubelet/pods/ae6cb076-b837-4671-83ba-43d479b94443/volumes" Dec 27 09:18:40 crc kubenswrapper[4934]: I1227 09:18:40.181521 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jqg5s" podUID="409de99a-6cd8-4574-8727-d59a71321d8d" containerName="registry-server" probeResult="failure" output=< Dec 27 09:18:40 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 09:18:40 crc kubenswrapper[4934]: > Dec 27 09:18:46 crc kubenswrapper[4934]: I1227 09:18:46.467983 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:18:46 crc kubenswrapper[4934]: E1227 09:18:46.469658 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:18:48 crc kubenswrapper[4934]: I1227 09:18:48.872933 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jqg5s" Dec 27 09:18:48 crc kubenswrapper[4934]: I1227 09:18:48.946745 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jqg5s" Dec 27 09:18:49 crc kubenswrapper[4934]: I1227 09:18:49.322941 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jqg5s"] Dec 27 09:18:50 crc kubenswrapper[4934]: I1227 09:18:50.378896 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jqg5s" podUID="409de99a-6cd8-4574-8727-d59a71321d8d" containerName="registry-server" containerID="cri-o://0db1a02b1ca44a9e50336ee70f02ff1246ea0200ffddfd7859d1b6a08b86ca36" gracePeriod=2 Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.015174 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jqg5s" Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.116901 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/409de99a-6cd8-4574-8727-d59a71321d8d-catalog-content\") pod \"409de99a-6cd8-4574-8727-d59a71321d8d\" (UID: \"409de99a-6cd8-4574-8727-d59a71321d8d\") " Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.116981 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xtxb\" (UniqueName: \"kubernetes.io/projected/409de99a-6cd8-4574-8727-d59a71321d8d-kube-api-access-7xtxb\") pod \"409de99a-6cd8-4574-8727-d59a71321d8d\" (UID: \"409de99a-6cd8-4574-8727-d59a71321d8d\") " Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.117455 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/409de99a-6cd8-4574-8727-d59a71321d8d-utilities\") pod \"409de99a-6cd8-4574-8727-d59a71321d8d\" (UID: \"409de99a-6cd8-4574-8727-d59a71321d8d\") " Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.118122 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/409de99a-6cd8-4574-8727-d59a71321d8d-utilities" (OuterVolumeSpecName: "utilities") pod "409de99a-6cd8-4574-8727-d59a71321d8d" (UID: "409de99a-6cd8-4574-8727-d59a71321d8d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.122628 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/409de99a-6cd8-4574-8727-d59a71321d8d-kube-api-access-7xtxb" (OuterVolumeSpecName: "kube-api-access-7xtxb") pod "409de99a-6cd8-4574-8727-d59a71321d8d" (UID: "409de99a-6cd8-4574-8727-d59a71321d8d"). InnerVolumeSpecName "kube-api-access-7xtxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.224813 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/409de99a-6cd8-4574-8727-d59a71321d8d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "409de99a-6cd8-4574-8727-d59a71321d8d" (UID: "409de99a-6cd8-4574-8727-d59a71321d8d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.227577 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/409de99a-6cd8-4574-8727-d59a71321d8d-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.227630 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/409de99a-6cd8-4574-8727-d59a71321d8d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.227654 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xtxb\" (UniqueName: \"kubernetes.io/projected/409de99a-6cd8-4574-8727-d59a71321d8d-kube-api-access-7xtxb\") on node \"crc\" DevicePath \"\"" Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.397191 4934 generic.go:334] "Generic (PLEG): container finished" podID="409de99a-6cd8-4574-8727-d59a71321d8d" containerID="0db1a02b1ca44a9e50336ee70f02ff1246ea0200ffddfd7859d1b6a08b86ca36" exitCode=0 Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.397511 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jqg5s" event={"ID":"409de99a-6cd8-4574-8727-d59a71321d8d","Type":"ContainerDied","Data":"0db1a02b1ca44a9e50336ee70f02ff1246ea0200ffddfd7859d1b6a08b86ca36"} Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.397544 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jqg5s" event={"ID":"409de99a-6cd8-4574-8727-d59a71321d8d","Type":"ContainerDied","Data":"f574bec91d60d052207b4cef6db920983b94d5534d7d3d50b3cd41540c07bb99"} Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.397566 4934 scope.go:117] "RemoveContainer" containerID="0db1a02b1ca44a9e50336ee70f02ff1246ea0200ffddfd7859d1b6a08b86ca36" Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.397612 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jqg5s" Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.442978 4934 scope.go:117] "RemoveContainer" containerID="bac3b1750a49cc4d2e80862c26063bc87029bd565f2ba33e23477eb00d27e221" Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.463678 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jqg5s"] Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.485284 4934 scope.go:117] "RemoveContainer" containerID="55dcd3da63e6bba81b9e89f34fbebadf34ab438239d03d13e61132eaf2f89ecc" Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.497722 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jqg5s"] Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.551797 4934 scope.go:117] "RemoveContainer" containerID="0db1a02b1ca44a9e50336ee70f02ff1246ea0200ffddfd7859d1b6a08b86ca36" Dec 27 09:18:51 crc kubenswrapper[4934]: E1227 09:18:51.552298 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0db1a02b1ca44a9e50336ee70f02ff1246ea0200ffddfd7859d1b6a08b86ca36\": container with ID starting with 0db1a02b1ca44a9e50336ee70f02ff1246ea0200ffddfd7859d1b6a08b86ca36 not found: ID does not exist" containerID="0db1a02b1ca44a9e50336ee70f02ff1246ea0200ffddfd7859d1b6a08b86ca36" Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.552330 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0db1a02b1ca44a9e50336ee70f02ff1246ea0200ffddfd7859d1b6a08b86ca36"} err="failed to get container status \"0db1a02b1ca44a9e50336ee70f02ff1246ea0200ffddfd7859d1b6a08b86ca36\": rpc error: code = NotFound desc = could not find container \"0db1a02b1ca44a9e50336ee70f02ff1246ea0200ffddfd7859d1b6a08b86ca36\": container with ID starting with 0db1a02b1ca44a9e50336ee70f02ff1246ea0200ffddfd7859d1b6a08b86ca36 not found: ID does not exist" Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.552353 4934 scope.go:117] "RemoveContainer" containerID="bac3b1750a49cc4d2e80862c26063bc87029bd565f2ba33e23477eb00d27e221" Dec 27 09:18:51 crc kubenswrapper[4934]: E1227 09:18:51.552866 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bac3b1750a49cc4d2e80862c26063bc87029bd565f2ba33e23477eb00d27e221\": container with ID starting with bac3b1750a49cc4d2e80862c26063bc87029bd565f2ba33e23477eb00d27e221 not found: ID does not exist" containerID="bac3b1750a49cc4d2e80862c26063bc87029bd565f2ba33e23477eb00d27e221" Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.552895 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bac3b1750a49cc4d2e80862c26063bc87029bd565f2ba33e23477eb00d27e221"} err="failed to get container status \"bac3b1750a49cc4d2e80862c26063bc87029bd565f2ba33e23477eb00d27e221\": rpc error: code = NotFound desc = could not find container \"bac3b1750a49cc4d2e80862c26063bc87029bd565f2ba33e23477eb00d27e221\": container with ID starting with bac3b1750a49cc4d2e80862c26063bc87029bd565f2ba33e23477eb00d27e221 not found: ID does not exist" Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.552914 4934 scope.go:117] "RemoveContainer" containerID="55dcd3da63e6bba81b9e89f34fbebadf34ab438239d03d13e61132eaf2f89ecc" Dec 27 09:18:51 crc kubenswrapper[4934]: E1227 09:18:51.553435 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55dcd3da63e6bba81b9e89f34fbebadf34ab438239d03d13e61132eaf2f89ecc\": container with ID starting with 55dcd3da63e6bba81b9e89f34fbebadf34ab438239d03d13e61132eaf2f89ecc not found: ID does not exist" containerID="55dcd3da63e6bba81b9e89f34fbebadf34ab438239d03d13e61132eaf2f89ecc" Dec 27 09:18:51 crc kubenswrapper[4934]: I1227 09:18:51.553604 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55dcd3da63e6bba81b9e89f34fbebadf34ab438239d03d13e61132eaf2f89ecc"} err="failed to get container status \"55dcd3da63e6bba81b9e89f34fbebadf34ab438239d03d13e61132eaf2f89ecc\": rpc error: code = NotFound desc = could not find container \"55dcd3da63e6bba81b9e89f34fbebadf34ab438239d03d13e61132eaf2f89ecc\": container with ID starting with 55dcd3da63e6bba81b9e89f34fbebadf34ab438239d03d13e61132eaf2f89ecc not found: ID does not exist" Dec 27 09:18:53 crc kubenswrapper[4934]: I1227 09:18:53.486647 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="409de99a-6cd8-4574-8727-d59a71321d8d" path="/var/lib/kubelet/pods/409de99a-6cd8-4574-8727-d59a71321d8d/volumes" Dec 27 09:18:57 crc kubenswrapper[4934]: I1227 09:18:57.467978 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:18:57 crc kubenswrapper[4934]: E1227 09:18:57.468787 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:19:08 crc kubenswrapper[4934]: I1227 09:19:08.467883 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:19:08 crc kubenswrapper[4934]: E1227 09:19:08.469337 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:19:12 crc kubenswrapper[4934]: I1227 09:19:12.642897 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="ca95a4ef-282d-4c26-983f-d7d00f90443c" containerName="galera" probeResult="failure" output="command timed out" Dec 27 09:19:22 crc kubenswrapper[4934]: I1227 09:19:22.467545 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:19:22 crc kubenswrapper[4934]: E1227 09:19:22.468325 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:19:34 crc kubenswrapper[4934]: I1227 09:19:34.468704 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:19:34 crc kubenswrapper[4934]: E1227 09:19:34.470332 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:19:46 crc kubenswrapper[4934]: I1227 09:19:46.468360 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:19:46 crc kubenswrapper[4934]: E1227 09:19:46.469430 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:20:01 crc kubenswrapper[4934]: I1227 09:20:01.469216 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:20:01 crc kubenswrapper[4934]: E1227 09:20:01.470220 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:20:12 crc kubenswrapper[4934]: I1227 09:20:12.467930 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:20:12 crc kubenswrapper[4934]: E1227 09:20:12.469305 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:20:27 crc kubenswrapper[4934]: I1227 09:20:27.469002 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:20:27 crc kubenswrapper[4934]: E1227 09:20:27.471305 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:20:39 crc kubenswrapper[4934]: I1227 09:20:39.476263 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:20:39 crc kubenswrapper[4934]: E1227 09:20:39.477048 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:20:45 crc kubenswrapper[4934]: I1227 09:20:45.937625 4934 generic.go:334] "Generic (PLEG): container finished" podID="d38686ee-0d94-4694-845a-e1ee5971af6f" containerID="087703ddc9f57a6c49b83599dfc1da1aa1da6ca6d5fd60592955500f5be1ab77" exitCode=1 Dec 27 09:20:45 crc kubenswrapper[4934]: I1227 09:20:45.937704 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d38686ee-0d94-4694-845a-e1ee5971af6f","Type":"ContainerDied","Data":"087703ddc9f57a6c49b83599dfc1da1aa1da6ca6d5fd60592955500f5be1ab77"} Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.446446 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.554076 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d38686ee-0d94-4694-845a-e1ee5971af6f-openstack-config\") pod \"d38686ee-0d94-4694-845a-e1ee5971af6f\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.555387 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d38686ee-0d94-4694-845a-e1ee5971af6f-ca-certs\") pod \"d38686ee-0d94-4694-845a-e1ee5971af6f\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.555496 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"d38686ee-0d94-4694-845a-e1ee5971af6f\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.555586 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d38686ee-0d94-4694-845a-e1ee5971af6f-config-data\") pod \"d38686ee-0d94-4694-845a-e1ee5971af6f\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.555722 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d38686ee-0d94-4694-845a-e1ee5971af6f-test-operator-ephemeral-workdir\") pod \"d38686ee-0d94-4694-845a-e1ee5971af6f\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.555821 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d38686ee-0d94-4694-845a-e1ee5971af6f-openstack-config-secret\") pod \"d38686ee-0d94-4694-845a-e1ee5971af6f\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.556121 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d38686ee-0d94-4694-845a-e1ee5971af6f-ssh-key\") pod \"d38686ee-0d94-4694-845a-e1ee5971af6f\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.556294 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d38686ee-0d94-4694-845a-e1ee5971af6f-test-operator-ephemeral-temporary\") pod \"d38686ee-0d94-4694-845a-e1ee5971af6f\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.556376 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2pr9\" (UniqueName: \"kubernetes.io/projected/d38686ee-0d94-4694-845a-e1ee5971af6f-kube-api-access-h2pr9\") pod \"d38686ee-0d94-4694-845a-e1ee5971af6f\" (UID: \"d38686ee-0d94-4694-845a-e1ee5971af6f\") " Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.556643 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d38686ee-0d94-4694-845a-e1ee5971af6f-config-data" (OuterVolumeSpecName: "config-data") pod "d38686ee-0d94-4694-845a-e1ee5971af6f" (UID: "d38686ee-0d94-4694-845a-e1ee5971af6f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.557912 4934 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d38686ee-0d94-4694-845a-e1ee5971af6f-config-data\") on node \"crc\" DevicePath \"\"" Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.558464 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d38686ee-0d94-4694-845a-e1ee5971af6f-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "d38686ee-0d94-4694-845a-e1ee5971af6f" (UID: "d38686ee-0d94-4694-845a-e1ee5971af6f"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.561909 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d38686ee-0d94-4694-845a-e1ee5971af6f-kube-api-access-h2pr9" (OuterVolumeSpecName: "kube-api-access-h2pr9") pod "d38686ee-0d94-4694-845a-e1ee5971af6f" (UID: "d38686ee-0d94-4694-845a-e1ee5971af6f"). InnerVolumeSpecName "kube-api-access-h2pr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.563786 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d38686ee-0d94-4694-845a-e1ee5971af6f-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "d38686ee-0d94-4694-845a-e1ee5971af6f" (UID: "d38686ee-0d94-4694-845a-e1ee5971af6f"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.566647 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "test-operator-logs") pod "d38686ee-0d94-4694-845a-e1ee5971af6f" (UID: "d38686ee-0d94-4694-845a-e1ee5971af6f"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.594369 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d38686ee-0d94-4694-845a-e1ee5971af6f-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "d38686ee-0d94-4694-845a-e1ee5971af6f" (UID: "d38686ee-0d94-4694-845a-e1ee5971af6f"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.600642 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d38686ee-0d94-4694-845a-e1ee5971af6f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d38686ee-0d94-4694-845a-e1ee5971af6f" (UID: "d38686ee-0d94-4694-845a-e1ee5971af6f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.602060 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d38686ee-0d94-4694-845a-e1ee5971af6f-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "d38686ee-0d94-4694-845a-e1ee5971af6f" (UID: "d38686ee-0d94-4694-845a-e1ee5971af6f"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.620440 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d38686ee-0d94-4694-845a-e1ee5971af6f-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "d38686ee-0d94-4694-845a-e1ee5971af6f" (UID: "d38686ee-0d94-4694-845a-e1ee5971af6f"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.659925 4934 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/d38686ee-0d94-4694-845a-e1ee5971af6f-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.659960 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2pr9\" (UniqueName: \"kubernetes.io/projected/d38686ee-0d94-4694-845a-e1ee5971af6f-kube-api-access-h2pr9\") on node \"crc\" DevicePath \"\"" Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.659971 4934 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d38686ee-0d94-4694-845a-e1ee5971af6f-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.659981 4934 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/d38686ee-0d94-4694-845a-e1ee5971af6f-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.660002 4934 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.660012 4934 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/d38686ee-0d94-4694-845a-e1ee5971af6f-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.660021 4934 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d38686ee-0d94-4694-845a-e1ee5971af6f-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.660029 4934 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d38686ee-0d94-4694-845a-e1ee5971af6f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.715444 4934 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.763627 4934 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.963326 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"d38686ee-0d94-4694-845a-e1ee5971af6f","Type":"ContainerDied","Data":"3a51ec0f1172ba43a52c38f4520a3082598a868ea2e3498e9e313f13152a967a"} Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.963665 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a51ec0f1172ba43a52c38f4520a3082598a868ea2e3498e9e313f13152a967a" Dec 27 09:20:47 crc kubenswrapper[4934]: I1227 09:20:47.963871 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.426834 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 27 09:20:53 crc kubenswrapper[4934]: E1227 09:20:53.428463 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="409de99a-6cd8-4574-8727-d59a71321d8d" containerName="extract-content" Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.428510 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="409de99a-6cd8-4574-8727-d59a71321d8d" containerName="extract-content" Dec 27 09:20:53 crc kubenswrapper[4934]: E1227 09:20:53.428560 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae6cb076-b837-4671-83ba-43d479b94443" containerName="registry-server" Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.428579 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae6cb076-b837-4671-83ba-43d479b94443" containerName="registry-server" Dec 27 09:20:53 crc kubenswrapper[4934]: E1227 09:20:53.428653 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="409de99a-6cd8-4574-8727-d59a71321d8d" containerName="registry-server" Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.428671 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="409de99a-6cd8-4574-8727-d59a71321d8d" containerName="registry-server" Dec 27 09:20:53 crc kubenswrapper[4934]: E1227 09:20:53.428716 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="409de99a-6cd8-4574-8727-d59a71321d8d" containerName="extract-utilities" Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.428734 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="409de99a-6cd8-4574-8727-d59a71321d8d" containerName="extract-utilities" Dec 27 09:20:53 crc kubenswrapper[4934]: E1227 09:20:53.428773 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae6cb076-b837-4671-83ba-43d479b94443" containerName="extract-content" Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.428787 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae6cb076-b837-4671-83ba-43d479b94443" containerName="extract-content" Dec 27 09:20:53 crc kubenswrapper[4934]: E1227 09:20:53.428846 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae6cb076-b837-4671-83ba-43d479b94443" containerName="extract-utilities" Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.428859 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae6cb076-b837-4671-83ba-43d479b94443" containerName="extract-utilities" Dec 27 09:20:53 crc kubenswrapper[4934]: E1227 09:20:53.428901 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d38686ee-0d94-4694-845a-e1ee5971af6f" containerName="tempest-tests-tempest-tests-runner" Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.428915 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="d38686ee-0d94-4694-845a-e1ee5971af6f" containerName="tempest-tests-tempest-tests-runner" Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.429624 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae6cb076-b837-4671-83ba-43d479b94443" containerName="registry-server" Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.429723 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="d38686ee-0d94-4694-845a-e1ee5971af6f" containerName="tempest-tests-tempest-tests-runner" Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.429756 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="409de99a-6cd8-4574-8727-d59a71321d8d" containerName="registry-server" Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.431518 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.435022 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-vjwww" Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.438672 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.605926 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2sqj\" (UniqueName: \"kubernetes.io/projected/aa0bb6c7-0651-4833-b1e0-364be9987257-kube-api-access-h2sqj\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"aa0bb6c7-0651-4833-b1e0-364be9987257\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.605984 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"aa0bb6c7-0651-4833-b1e0-364be9987257\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.710066 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2sqj\" (UniqueName: \"kubernetes.io/projected/aa0bb6c7-0651-4833-b1e0-364be9987257-kube-api-access-h2sqj\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"aa0bb6c7-0651-4833-b1e0-364be9987257\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.710175 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"aa0bb6c7-0651-4833-b1e0-364be9987257\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.711351 4934 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"aa0bb6c7-0651-4833-b1e0-364be9987257\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.728845 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2sqj\" (UniqueName: \"kubernetes.io/projected/aa0bb6c7-0651-4833-b1e0-364be9987257-kube-api-access-h2sqj\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"aa0bb6c7-0651-4833-b1e0-364be9987257\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.746966 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"aa0bb6c7-0651-4833-b1e0-364be9987257\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 27 09:20:53 crc kubenswrapper[4934]: I1227 09:20:53.770798 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 27 09:20:54 crc kubenswrapper[4934]: I1227 09:20:54.265450 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 27 09:20:54 crc kubenswrapper[4934]: I1227 09:20:54.468312 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:20:54 crc kubenswrapper[4934]: E1227 09:20:54.469043 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:20:55 crc kubenswrapper[4934]: I1227 09:20:55.061625 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"aa0bb6c7-0651-4833-b1e0-364be9987257","Type":"ContainerStarted","Data":"b416f495f88ffc45667d6f2b402dc62311ba906daa2289b3f20429b7a9ede3ec"} Dec 27 09:20:56 crc kubenswrapper[4934]: I1227 09:20:56.072948 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"aa0bb6c7-0651-4833-b1e0-364be9987257","Type":"ContainerStarted","Data":"36734d0f602a7eec6aec379373d6da22c9fa629f47ed58a0cd54adc9e425ae84"} Dec 27 09:20:56 crc kubenswrapper[4934]: I1227 09:20:56.091103 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.9791193969999998 podStartE2EDuration="3.091058304s" podCreationTimestamp="2025-12-27 09:20:53 +0000 UTC" firstStartedPulling="2025-12-27 09:20:54.274763437 +0000 UTC m=+5915.095204031" lastFinishedPulling="2025-12-27 09:20:55.386702334 +0000 UTC m=+5916.207142938" observedRunningTime="2025-12-27 09:20:56.090231224 +0000 UTC m=+5916.910671818" watchObservedRunningTime="2025-12-27 09:20:56.091058304 +0000 UTC m=+5916.911498898" Dec 27 09:21:05 crc kubenswrapper[4934]: I1227 09:21:05.469458 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:21:05 crc kubenswrapper[4934]: E1227 09:21:05.471062 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:21:16 crc kubenswrapper[4934]: I1227 09:21:16.468580 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:21:16 crc kubenswrapper[4934]: E1227 09:21:16.469735 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:21:25 crc kubenswrapper[4934]: I1227 09:21:25.770289 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-c8rzb/must-gather-f4p5q"] Dec 27 09:21:25 crc kubenswrapper[4934]: I1227 09:21:25.774276 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8rzb/must-gather-f4p5q" Dec 27 09:21:25 crc kubenswrapper[4934]: I1227 09:21:25.777350 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-c8rzb"/"default-dockercfg-bsp98" Dec 27 09:21:25 crc kubenswrapper[4934]: I1227 09:21:25.779015 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-c8rzb"/"openshift-service-ca.crt" Dec 27 09:21:25 crc kubenswrapper[4934]: I1227 09:21:25.795173 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-c8rzb/must-gather-f4p5q"] Dec 27 09:21:25 crc kubenswrapper[4934]: I1227 09:21:25.801001 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-c8rzb"/"kube-root-ca.crt" Dec 27 09:21:25 crc kubenswrapper[4934]: I1227 09:21:25.886393 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdfn7\" (UniqueName: \"kubernetes.io/projected/e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f-kube-api-access-zdfn7\") pod \"must-gather-f4p5q\" (UID: \"e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f\") " pod="openshift-must-gather-c8rzb/must-gather-f4p5q" Dec 27 09:21:25 crc kubenswrapper[4934]: I1227 09:21:25.886618 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f-must-gather-output\") pod \"must-gather-f4p5q\" (UID: \"e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f\") " pod="openshift-must-gather-c8rzb/must-gather-f4p5q" Dec 27 09:21:25 crc kubenswrapper[4934]: I1227 09:21:25.989297 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f-must-gather-output\") pod \"must-gather-f4p5q\" (UID: \"e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f\") " pod="openshift-must-gather-c8rzb/must-gather-f4p5q" Dec 27 09:21:25 crc kubenswrapper[4934]: I1227 09:21:25.989505 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdfn7\" (UniqueName: \"kubernetes.io/projected/e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f-kube-api-access-zdfn7\") pod \"must-gather-f4p5q\" (UID: \"e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f\") " pod="openshift-must-gather-c8rzb/must-gather-f4p5q" Dec 27 09:21:25 crc kubenswrapper[4934]: I1227 09:21:25.989725 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f-must-gather-output\") pod \"must-gather-f4p5q\" (UID: \"e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f\") " pod="openshift-must-gather-c8rzb/must-gather-f4p5q" Dec 27 09:21:26 crc kubenswrapper[4934]: I1227 09:21:26.140077 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdfn7\" (UniqueName: \"kubernetes.io/projected/e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f-kube-api-access-zdfn7\") pod \"must-gather-f4p5q\" (UID: \"e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f\") " pod="openshift-must-gather-c8rzb/must-gather-f4p5q" Dec 27 09:21:26 crc kubenswrapper[4934]: I1227 09:21:26.403296 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8rzb/must-gather-f4p5q" Dec 27 09:21:26 crc kubenswrapper[4934]: I1227 09:21:26.890023 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-c8rzb/must-gather-f4p5q"] Dec 27 09:21:27 crc kubenswrapper[4934]: I1227 09:21:27.540269 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8rzb/must-gather-f4p5q" event={"ID":"e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f","Type":"ContainerStarted","Data":"2ef3394026860cddc3c9470407651b34df69080f83cdd32f5a933c4d73e8749d"} Dec 27 09:21:31 crc kubenswrapper[4934]: I1227 09:21:31.468840 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:21:31 crc kubenswrapper[4934]: E1227 09:21:31.469521 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:21:34 crc kubenswrapper[4934]: I1227 09:21:34.659550 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8rzb/must-gather-f4p5q" event={"ID":"e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f","Type":"ContainerStarted","Data":"44b230a08f0db8ff213b6452d3159f62e7835d4f9033dfd4649d72ae1396c7d9"} Dec 27 09:21:35 crc kubenswrapper[4934]: I1227 09:21:35.676685 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8rzb/must-gather-f4p5q" event={"ID":"e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f","Type":"ContainerStarted","Data":"0ce63c7ad4fb139ab4564105fa1322b7292f8c7a0f12262d42f4fcab821aa1ec"} Dec 27 09:21:35 crc kubenswrapper[4934]: I1227 09:21:35.713670 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-c8rzb/must-gather-f4p5q" podStartSLOduration=3.3825863910000002 podStartE2EDuration="10.713650479s" podCreationTimestamp="2025-12-27 09:21:25 +0000 UTC" firstStartedPulling="2025-12-27 09:21:26.888067177 +0000 UTC m=+5947.708507771" lastFinishedPulling="2025-12-27 09:21:34.219131265 +0000 UTC m=+5955.039571859" observedRunningTime="2025-12-27 09:21:35.690105876 +0000 UTC m=+5956.510546480" watchObservedRunningTime="2025-12-27 09:21:35.713650479 +0000 UTC m=+5956.534091073" Dec 27 09:21:39 crc kubenswrapper[4934]: I1227 09:21:39.384590 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-c8rzb/crc-debug-7brmx"] Dec 27 09:21:39 crc kubenswrapper[4934]: I1227 09:21:39.388679 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8rzb/crc-debug-7brmx" Dec 27 09:21:39 crc kubenswrapper[4934]: I1227 09:21:39.530222 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fee7329-21ad-439a-9500-bbe4d4220c93-host\") pod \"crc-debug-7brmx\" (UID: \"9fee7329-21ad-439a-9500-bbe4d4220c93\") " pod="openshift-must-gather-c8rzb/crc-debug-7brmx" Dec 27 09:21:39 crc kubenswrapper[4934]: I1227 09:21:39.530281 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbmkx\" (UniqueName: \"kubernetes.io/projected/9fee7329-21ad-439a-9500-bbe4d4220c93-kube-api-access-rbmkx\") pod \"crc-debug-7brmx\" (UID: \"9fee7329-21ad-439a-9500-bbe4d4220c93\") " pod="openshift-must-gather-c8rzb/crc-debug-7brmx" Dec 27 09:21:39 crc kubenswrapper[4934]: I1227 09:21:39.634011 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fee7329-21ad-439a-9500-bbe4d4220c93-host\") pod \"crc-debug-7brmx\" (UID: \"9fee7329-21ad-439a-9500-bbe4d4220c93\") " pod="openshift-must-gather-c8rzb/crc-debug-7brmx" Dec 27 09:21:39 crc kubenswrapper[4934]: I1227 09:21:39.634095 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbmkx\" (UniqueName: \"kubernetes.io/projected/9fee7329-21ad-439a-9500-bbe4d4220c93-kube-api-access-rbmkx\") pod \"crc-debug-7brmx\" (UID: \"9fee7329-21ad-439a-9500-bbe4d4220c93\") " pod="openshift-must-gather-c8rzb/crc-debug-7brmx" Dec 27 09:21:39 crc kubenswrapper[4934]: I1227 09:21:39.634292 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fee7329-21ad-439a-9500-bbe4d4220c93-host\") pod \"crc-debug-7brmx\" (UID: \"9fee7329-21ad-439a-9500-bbe4d4220c93\") " pod="openshift-must-gather-c8rzb/crc-debug-7brmx" Dec 27 09:21:39 crc kubenswrapper[4934]: I1227 09:21:39.656977 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbmkx\" (UniqueName: \"kubernetes.io/projected/9fee7329-21ad-439a-9500-bbe4d4220c93-kube-api-access-rbmkx\") pod \"crc-debug-7brmx\" (UID: \"9fee7329-21ad-439a-9500-bbe4d4220c93\") " pod="openshift-must-gather-c8rzb/crc-debug-7brmx" Dec 27 09:21:39 crc kubenswrapper[4934]: I1227 09:21:39.718405 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8rzb/crc-debug-7brmx" Dec 27 09:21:39 crc kubenswrapper[4934]: W1227 09:21:39.763207 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fee7329_21ad_439a_9500_bbe4d4220c93.slice/crio-149e0ba94540c9776f0173cb7e68a9850a615739eebcd6fa1d50e7a0a1fd2992 WatchSource:0}: Error finding container 149e0ba94540c9776f0173cb7e68a9850a615739eebcd6fa1d50e7a0a1fd2992: Status 404 returned error can't find the container with id 149e0ba94540c9776f0173cb7e68a9850a615739eebcd6fa1d50e7a0a1fd2992 Dec 27 09:21:40 crc kubenswrapper[4934]: I1227 09:21:40.727450 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8rzb/crc-debug-7brmx" event={"ID":"9fee7329-21ad-439a-9500-bbe4d4220c93","Type":"ContainerStarted","Data":"149e0ba94540c9776f0173cb7e68a9850a615739eebcd6fa1d50e7a0a1fd2992"} Dec 27 09:21:46 crc kubenswrapper[4934]: I1227 09:21:46.467879 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:21:46 crc kubenswrapper[4934]: E1227 09:21:46.468796 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:21:50 crc kubenswrapper[4934]: I1227 09:21:50.853030 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8rzb/crc-debug-7brmx" event={"ID":"9fee7329-21ad-439a-9500-bbe4d4220c93","Type":"ContainerStarted","Data":"4e2cd24c5bb202389015961ea4fc46d3be488f2e28def4a78ade1df5822ae4b4"} Dec 27 09:21:50 crc kubenswrapper[4934]: I1227 09:21:50.867608 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-c8rzb/crc-debug-7brmx" podStartSLOduration=1.5474191959999999 podStartE2EDuration="11.867590194s" podCreationTimestamp="2025-12-27 09:21:39 +0000 UTC" firstStartedPulling="2025-12-27 09:21:39.766764308 +0000 UTC m=+5960.587204902" lastFinishedPulling="2025-12-27 09:21:50.086935306 +0000 UTC m=+5970.907375900" observedRunningTime="2025-12-27 09:21:50.865678086 +0000 UTC m=+5971.686118720" watchObservedRunningTime="2025-12-27 09:21:50.867590194 +0000 UTC m=+5971.688030788" Dec 27 09:21:58 crc kubenswrapper[4934]: I1227 09:21:58.468695 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:21:58 crc kubenswrapper[4934]: E1227 09:21:58.470103 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:22:10 crc kubenswrapper[4934]: I1227 09:22:10.468488 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:22:10 crc kubenswrapper[4934]: E1227 09:22:10.469421 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:22:22 crc kubenswrapper[4934]: I1227 09:22:22.468431 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:22:22 crc kubenswrapper[4934]: E1227 09:22:22.469258 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:22:35 crc kubenswrapper[4934]: I1227 09:22:35.467297 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:22:35 crc kubenswrapper[4934]: E1227 09:22:35.469625 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:22:41 crc kubenswrapper[4934]: I1227 09:22:41.476282 4934 generic.go:334] "Generic (PLEG): container finished" podID="9fee7329-21ad-439a-9500-bbe4d4220c93" containerID="4e2cd24c5bb202389015961ea4fc46d3be488f2e28def4a78ade1df5822ae4b4" exitCode=0 Dec 27 09:22:41 crc kubenswrapper[4934]: I1227 09:22:41.487457 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8rzb/crc-debug-7brmx" event={"ID":"9fee7329-21ad-439a-9500-bbe4d4220c93","Type":"ContainerDied","Data":"4e2cd24c5bb202389015961ea4fc46d3be488f2e28def4a78ade1df5822ae4b4"} Dec 27 09:22:42 crc kubenswrapper[4934]: I1227 09:22:42.620799 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8rzb/crc-debug-7brmx" Dec 27 09:22:42 crc kubenswrapper[4934]: I1227 09:22:42.663380 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-c8rzb/crc-debug-7brmx"] Dec 27 09:22:42 crc kubenswrapper[4934]: I1227 09:22:42.675420 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-c8rzb/crc-debug-7brmx"] Dec 27 09:22:42 crc kubenswrapper[4934]: I1227 09:22:42.733144 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fee7329-21ad-439a-9500-bbe4d4220c93-host\") pod \"9fee7329-21ad-439a-9500-bbe4d4220c93\" (UID: \"9fee7329-21ad-439a-9500-bbe4d4220c93\") " Dec 27 09:22:42 crc kubenswrapper[4934]: I1227 09:22:42.733242 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbmkx\" (UniqueName: \"kubernetes.io/projected/9fee7329-21ad-439a-9500-bbe4d4220c93-kube-api-access-rbmkx\") pod \"9fee7329-21ad-439a-9500-bbe4d4220c93\" (UID: \"9fee7329-21ad-439a-9500-bbe4d4220c93\") " Dec 27 09:22:42 crc kubenswrapper[4934]: I1227 09:22:42.733259 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9fee7329-21ad-439a-9500-bbe4d4220c93-host" (OuterVolumeSpecName: "host") pod "9fee7329-21ad-439a-9500-bbe4d4220c93" (UID: "9fee7329-21ad-439a-9500-bbe4d4220c93"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 09:22:42 crc kubenswrapper[4934]: I1227 09:22:42.733882 4934 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fee7329-21ad-439a-9500-bbe4d4220c93-host\") on node \"crc\" DevicePath \"\"" Dec 27 09:22:42 crc kubenswrapper[4934]: I1227 09:22:42.741830 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fee7329-21ad-439a-9500-bbe4d4220c93-kube-api-access-rbmkx" (OuterVolumeSpecName: "kube-api-access-rbmkx") pod "9fee7329-21ad-439a-9500-bbe4d4220c93" (UID: "9fee7329-21ad-439a-9500-bbe4d4220c93"). InnerVolumeSpecName "kube-api-access-rbmkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:22:42 crc kubenswrapper[4934]: I1227 09:22:42.836025 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbmkx\" (UniqueName: \"kubernetes.io/projected/9fee7329-21ad-439a-9500-bbe4d4220c93-kube-api-access-rbmkx\") on node \"crc\" DevicePath \"\"" Dec 27 09:22:44 crc kubenswrapper[4934]: I1227 09:22:43.483734 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fee7329-21ad-439a-9500-bbe4d4220c93" path="/var/lib/kubelet/pods/9fee7329-21ad-439a-9500-bbe4d4220c93/volumes" Dec 27 09:22:44 crc kubenswrapper[4934]: I1227 09:22:43.502624 4934 scope.go:117] "RemoveContainer" containerID="4e2cd24c5bb202389015961ea4fc46d3be488f2e28def4a78ade1df5822ae4b4" Dec 27 09:22:44 crc kubenswrapper[4934]: I1227 09:22:43.502741 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8rzb/crc-debug-7brmx" Dec 27 09:22:44 crc kubenswrapper[4934]: I1227 09:22:44.765645 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-c8rzb/crc-debug-k4xbf"] Dec 27 09:22:44 crc kubenswrapper[4934]: E1227 09:22:44.766248 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fee7329-21ad-439a-9500-bbe4d4220c93" containerName="container-00" Dec 27 09:22:44 crc kubenswrapper[4934]: I1227 09:22:44.766266 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fee7329-21ad-439a-9500-bbe4d4220c93" containerName="container-00" Dec 27 09:22:44 crc kubenswrapper[4934]: I1227 09:22:44.766490 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fee7329-21ad-439a-9500-bbe4d4220c93" containerName="container-00" Dec 27 09:22:44 crc kubenswrapper[4934]: I1227 09:22:44.767326 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8rzb/crc-debug-k4xbf" Dec 27 09:22:44 crc kubenswrapper[4934]: I1227 09:22:44.898231 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w78l\" (UniqueName: \"kubernetes.io/projected/333879cd-8fde-4bd1-a996-69aec3798445-kube-api-access-4w78l\") pod \"crc-debug-k4xbf\" (UID: \"333879cd-8fde-4bd1-a996-69aec3798445\") " pod="openshift-must-gather-c8rzb/crc-debug-k4xbf" Dec 27 09:22:44 crc kubenswrapper[4934]: I1227 09:22:44.898330 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/333879cd-8fde-4bd1-a996-69aec3798445-host\") pod \"crc-debug-k4xbf\" (UID: \"333879cd-8fde-4bd1-a996-69aec3798445\") " pod="openshift-must-gather-c8rzb/crc-debug-k4xbf" Dec 27 09:22:45 crc kubenswrapper[4934]: I1227 09:22:45.000997 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w78l\" (UniqueName: \"kubernetes.io/projected/333879cd-8fde-4bd1-a996-69aec3798445-kube-api-access-4w78l\") pod \"crc-debug-k4xbf\" (UID: \"333879cd-8fde-4bd1-a996-69aec3798445\") " pod="openshift-must-gather-c8rzb/crc-debug-k4xbf" Dec 27 09:22:45 crc kubenswrapper[4934]: I1227 09:22:45.001098 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/333879cd-8fde-4bd1-a996-69aec3798445-host\") pod \"crc-debug-k4xbf\" (UID: \"333879cd-8fde-4bd1-a996-69aec3798445\") " pod="openshift-must-gather-c8rzb/crc-debug-k4xbf" Dec 27 09:22:45 crc kubenswrapper[4934]: I1227 09:22:45.001262 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/333879cd-8fde-4bd1-a996-69aec3798445-host\") pod \"crc-debug-k4xbf\" (UID: \"333879cd-8fde-4bd1-a996-69aec3798445\") " pod="openshift-must-gather-c8rzb/crc-debug-k4xbf" Dec 27 09:22:45 crc kubenswrapper[4934]: I1227 09:22:45.021719 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w78l\" (UniqueName: \"kubernetes.io/projected/333879cd-8fde-4bd1-a996-69aec3798445-kube-api-access-4w78l\") pod \"crc-debug-k4xbf\" (UID: \"333879cd-8fde-4bd1-a996-69aec3798445\") " pod="openshift-must-gather-c8rzb/crc-debug-k4xbf" Dec 27 09:22:45 crc kubenswrapper[4934]: I1227 09:22:45.093442 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8rzb/crc-debug-k4xbf" Dec 27 09:22:45 crc kubenswrapper[4934]: I1227 09:22:45.538537 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8rzb/crc-debug-k4xbf" event={"ID":"333879cd-8fde-4bd1-a996-69aec3798445","Type":"ContainerStarted","Data":"aa396c27848324ad149152cf42494d4e1b146e9be00c3264689f041563269213"} Dec 27 09:22:45 crc kubenswrapper[4934]: I1227 09:22:45.538605 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8rzb/crc-debug-k4xbf" event={"ID":"333879cd-8fde-4bd1-a996-69aec3798445","Type":"ContainerStarted","Data":"0a2adb85663fb5114912931115a9250c0cafe952dbfda35654eaec439c8d8e6e"} Dec 27 09:22:45 crc kubenswrapper[4934]: I1227 09:22:45.561384 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-c8rzb/crc-debug-k4xbf" podStartSLOduration=1.561365346 podStartE2EDuration="1.561365346s" podCreationTimestamp="2025-12-27 09:22:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-27 09:22:45.553750638 +0000 UTC m=+6026.374191232" watchObservedRunningTime="2025-12-27 09:22:45.561365346 +0000 UTC m=+6026.381805940" Dec 27 09:22:45 crc kubenswrapper[4934]: E1227 09:22:45.908063 4934 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod333879cd_8fde_4bd1_a996_69aec3798445.slice/crio-aa396c27848324ad149152cf42494d4e1b146e9be00c3264689f041563269213.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod333879cd_8fde_4bd1_a996_69aec3798445.slice/crio-conmon-aa396c27848324ad149152cf42494d4e1b146e9be00c3264689f041563269213.scope\": RecentStats: unable to find data in memory cache]" Dec 27 09:22:46 crc kubenswrapper[4934]: I1227 09:22:46.550426 4934 generic.go:334] "Generic (PLEG): container finished" podID="333879cd-8fde-4bd1-a996-69aec3798445" containerID="aa396c27848324ad149152cf42494d4e1b146e9be00c3264689f041563269213" exitCode=0 Dec 27 09:22:46 crc kubenswrapper[4934]: I1227 09:22:46.550677 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8rzb/crc-debug-k4xbf" event={"ID":"333879cd-8fde-4bd1-a996-69aec3798445","Type":"ContainerDied","Data":"aa396c27848324ad149152cf42494d4e1b146e9be00c3264689f041563269213"} Dec 27 09:22:47 crc kubenswrapper[4934]: I1227 09:22:47.683411 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8rzb/crc-debug-k4xbf" Dec 27 09:22:47 crc kubenswrapper[4934]: I1227 09:22:47.867171 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4w78l\" (UniqueName: \"kubernetes.io/projected/333879cd-8fde-4bd1-a996-69aec3798445-kube-api-access-4w78l\") pod \"333879cd-8fde-4bd1-a996-69aec3798445\" (UID: \"333879cd-8fde-4bd1-a996-69aec3798445\") " Dec 27 09:22:47 crc kubenswrapper[4934]: I1227 09:22:47.867649 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/333879cd-8fde-4bd1-a996-69aec3798445-host\") pod \"333879cd-8fde-4bd1-a996-69aec3798445\" (UID: \"333879cd-8fde-4bd1-a996-69aec3798445\") " Dec 27 09:22:47 crc kubenswrapper[4934]: I1227 09:22:47.868290 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/333879cd-8fde-4bd1-a996-69aec3798445-host" (OuterVolumeSpecName: "host") pod "333879cd-8fde-4bd1-a996-69aec3798445" (UID: "333879cd-8fde-4bd1-a996-69aec3798445"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 09:22:47 crc kubenswrapper[4934]: I1227 09:22:47.870570 4934 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/333879cd-8fde-4bd1-a996-69aec3798445-host\") on node \"crc\" DevicePath \"\"" Dec 27 09:22:47 crc kubenswrapper[4934]: I1227 09:22:47.876888 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/333879cd-8fde-4bd1-a996-69aec3798445-kube-api-access-4w78l" (OuterVolumeSpecName: "kube-api-access-4w78l") pod "333879cd-8fde-4bd1-a996-69aec3798445" (UID: "333879cd-8fde-4bd1-a996-69aec3798445"). InnerVolumeSpecName "kube-api-access-4w78l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:22:47 crc kubenswrapper[4934]: I1227 09:22:47.972629 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4w78l\" (UniqueName: \"kubernetes.io/projected/333879cd-8fde-4bd1-a996-69aec3798445-kube-api-access-4w78l\") on node \"crc\" DevicePath \"\"" Dec 27 09:22:48 crc kubenswrapper[4934]: I1227 09:22:48.137131 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-c8rzb/crc-debug-k4xbf"] Dec 27 09:22:48 crc kubenswrapper[4934]: I1227 09:22:48.154966 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-c8rzb/crc-debug-k4xbf"] Dec 27 09:22:48 crc kubenswrapper[4934]: I1227 09:22:48.581075 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a2adb85663fb5114912931115a9250c0cafe952dbfda35654eaec439c8d8e6e" Dec 27 09:22:48 crc kubenswrapper[4934]: I1227 09:22:48.581184 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8rzb/crc-debug-k4xbf" Dec 27 09:22:49 crc kubenswrapper[4934]: I1227 09:22:49.349789 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-c8rzb/crc-debug-sljrr"] Dec 27 09:22:49 crc kubenswrapper[4934]: E1227 09:22:49.352012 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="333879cd-8fde-4bd1-a996-69aec3798445" containerName="container-00" Dec 27 09:22:49 crc kubenswrapper[4934]: I1227 09:22:49.352059 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="333879cd-8fde-4bd1-a996-69aec3798445" containerName="container-00" Dec 27 09:22:49 crc kubenswrapper[4934]: I1227 09:22:49.352651 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="333879cd-8fde-4bd1-a996-69aec3798445" containerName="container-00" Dec 27 09:22:49 crc kubenswrapper[4934]: I1227 09:22:49.354064 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8rzb/crc-debug-sljrr" Dec 27 09:22:49 crc kubenswrapper[4934]: I1227 09:22:49.489603 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="333879cd-8fde-4bd1-a996-69aec3798445" path="/var/lib/kubelet/pods/333879cd-8fde-4bd1-a996-69aec3798445/volumes" Dec 27 09:22:49 crc kubenswrapper[4934]: I1227 09:22:49.510813 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee2dcc57-b167-4885-8a2a-b05979c842f7-host\") pod \"crc-debug-sljrr\" (UID: \"ee2dcc57-b167-4885-8a2a-b05979c842f7\") " pod="openshift-must-gather-c8rzb/crc-debug-sljrr" Dec 27 09:22:49 crc kubenswrapper[4934]: I1227 09:22:49.511254 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fr4ff\" (UniqueName: \"kubernetes.io/projected/ee2dcc57-b167-4885-8a2a-b05979c842f7-kube-api-access-fr4ff\") pod \"crc-debug-sljrr\" (UID: \"ee2dcc57-b167-4885-8a2a-b05979c842f7\") " pod="openshift-must-gather-c8rzb/crc-debug-sljrr" Dec 27 09:22:49 crc kubenswrapper[4934]: I1227 09:22:49.612933 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee2dcc57-b167-4885-8a2a-b05979c842f7-host\") pod \"crc-debug-sljrr\" (UID: \"ee2dcc57-b167-4885-8a2a-b05979c842f7\") " pod="openshift-must-gather-c8rzb/crc-debug-sljrr" Dec 27 09:22:49 crc kubenswrapper[4934]: I1227 09:22:49.613118 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee2dcc57-b167-4885-8a2a-b05979c842f7-host\") pod \"crc-debug-sljrr\" (UID: \"ee2dcc57-b167-4885-8a2a-b05979c842f7\") " pod="openshift-must-gather-c8rzb/crc-debug-sljrr" Dec 27 09:22:49 crc kubenswrapper[4934]: I1227 09:22:49.613671 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fr4ff\" (UniqueName: \"kubernetes.io/projected/ee2dcc57-b167-4885-8a2a-b05979c842f7-kube-api-access-fr4ff\") pod \"crc-debug-sljrr\" (UID: \"ee2dcc57-b167-4885-8a2a-b05979c842f7\") " pod="openshift-must-gather-c8rzb/crc-debug-sljrr" Dec 27 09:22:49 crc kubenswrapper[4934]: I1227 09:22:49.641070 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fr4ff\" (UniqueName: \"kubernetes.io/projected/ee2dcc57-b167-4885-8a2a-b05979c842f7-kube-api-access-fr4ff\") pod \"crc-debug-sljrr\" (UID: \"ee2dcc57-b167-4885-8a2a-b05979c842f7\") " pod="openshift-must-gather-c8rzb/crc-debug-sljrr" Dec 27 09:22:49 crc kubenswrapper[4934]: I1227 09:22:49.677913 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8rzb/crc-debug-sljrr" Dec 27 09:22:50 crc kubenswrapper[4934]: I1227 09:22:50.467823 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:22:50 crc kubenswrapper[4934]: I1227 09:22:50.613169 4934 generic.go:334] "Generic (PLEG): container finished" podID="ee2dcc57-b167-4885-8a2a-b05979c842f7" containerID="a9ed28976530e0e813009b36d70dcd5298fbfde0e17d06d1a99dece527c55182" exitCode=0 Dec 27 09:22:50 crc kubenswrapper[4934]: I1227 09:22:50.613235 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8rzb/crc-debug-sljrr" event={"ID":"ee2dcc57-b167-4885-8a2a-b05979c842f7","Type":"ContainerDied","Data":"a9ed28976530e0e813009b36d70dcd5298fbfde0e17d06d1a99dece527c55182"} Dec 27 09:22:50 crc kubenswrapper[4934]: I1227 09:22:50.613276 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8rzb/crc-debug-sljrr" event={"ID":"ee2dcc57-b167-4885-8a2a-b05979c842f7","Type":"ContainerStarted","Data":"fb35f06d2a1240e0504a3bea6b21d560b0e09b5fed98eb422540a18611551842"} Dec 27 09:22:50 crc kubenswrapper[4934]: I1227 09:22:50.684347 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-c8rzb/crc-debug-sljrr"] Dec 27 09:22:50 crc kubenswrapper[4934]: I1227 09:22:50.697691 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-c8rzb/crc-debug-sljrr"] Dec 27 09:22:51 crc kubenswrapper[4934]: I1227 09:22:51.639857 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"f3e74ce13c29f6896d91ffd6e9150b3e5f47b60d6ed52c857f27a4eb1e7598d0"} Dec 27 09:22:51 crc kubenswrapper[4934]: I1227 09:22:51.805488 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8rzb/crc-debug-sljrr" Dec 27 09:22:51 crc kubenswrapper[4934]: I1227 09:22:51.977654 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fr4ff\" (UniqueName: \"kubernetes.io/projected/ee2dcc57-b167-4885-8a2a-b05979c842f7-kube-api-access-fr4ff\") pod \"ee2dcc57-b167-4885-8a2a-b05979c842f7\" (UID: \"ee2dcc57-b167-4885-8a2a-b05979c842f7\") " Dec 27 09:22:51 crc kubenswrapper[4934]: I1227 09:22:51.977911 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee2dcc57-b167-4885-8a2a-b05979c842f7-host\") pod \"ee2dcc57-b167-4885-8a2a-b05979c842f7\" (UID: \"ee2dcc57-b167-4885-8a2a-b05979c842f7\") " Dec 27 09:22:51 crc kubenswrapper[4934]: I1227 09:22:51.977991 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee2dcc57-b167-4885-8a2a-b05979c842f7-host" (OuterVolumeSpecName: "host") pod "ee2dcc57-b167-4885-8a2a-b05979c842f7" (UID: "ee2dcc57-b167-4885-8a2a-b05979c842f7"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 27 09:22:51 crc kubenswrapper[4934]: I1227 09:22:51.978505 4934 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ee2dcc57-b167-4885-8a2a-b05979c842f7-host\") on node \"crc\" DevicePath \"\"" Dec 27 09:22:51 crc kubenswrapper[4934]: I1227 09:22:51.983837 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee2dcc57-b167-4885-8a2a-b05979c842f7-kube-api-access-fr4ff" (OuterVolumeSpecName: "kube-api-access-fr4ff") pod "ee2dcc57-b167-4885-8a2a-b05979c842f7" (UID: "ee2dcc57-b167-4885-8a2a-b05979c842f7"). InnerVolumeSpecName "kube-api-access-fr4ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:22:52 crc kubenswrapper[4934]: I1227 09:22:52.081343 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fr4ff\" (UniqueName: \"kubernetes.io/projected/ee2dcc57-b167-4885-8a2a-b05979c842f7-kube-api-access-fr4ff\") on node \"crc\" DevicePath \"\"" Dec 27 09:22:52 crc kubenswrapper[4934]: I1227 09:22:52.661902 4934 scope.go:117] "RemoveContainer" containerID="a9ed28976530e0e813009b36d70dcd5298fbfde0e17d06d1a99dece527c55182" Dec 27 09:22:52 crc kubenswrapper[4934]: I1227 09:22:52.661917 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8rzb/crc-debug-sljrr" Dec 27 09:22:53 crc kubenswrapper[4934]: I1227 09:22:53.478802 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee2dcc57-b167-4885-8a2a-b05979c842f7" path="/var/lib/kubelet/pods/ee2dcc57-b167-4885-8a2a-b05979c842f7/volumes" Dec 27 09:23:18 crc kubenswrapper[4934]: I1227 09:23:18.712928 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_ca41a5da-70d4-4668-ac75-978e192e46f4/aodh-api/0.log" Dec 27 09:23:18 crc kubenswrapper[4934]: I1227 09:23:18.933448 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_ca41a5da-70d4-4668-ac75-978e192e46f4/aodh-evaluator/0.log" Dec 27 09:23:18 crc kubenswrapper[4934]: I1227 09:23:18.940602 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_ca41a5da-70d4-4668-ac75-978e192e46f4/aodh-listener/0.log" Dec 27 09:23:19 crc kubenswrapper[4934]: I1227 09:23:19.015310 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_ca41a5da-70d4-4668-ac75-978e192e46f4/aodh-notifier/0.log" Dec 27 09:23:19 crc kubenswrapper[4934]: I1227 09:23:19.134917 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-76d79bd746-4mrv6_00fe7217-c021-410c-9fca-36b69e513e92/barbican-api-log/0.log" Dec 27 09:23:19 crc kubenswrapper[4934]: I1227 09:23:19.150438 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-76d79bd746-4mrv6_00fe7217-c021-410c-9fca-36b69e513e92/barbican-api/0.log" Dec 27 09:23:19 crc kubenswrapper[4934]: I1227 09:23:19.303430 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-654859b66d-km642_700034e0-f221-4bca-a973-a6d49c7dc406/barbican-keystone-listener/0.log" Dec 27 09:23:19 crc kubenswrapper[4934]: I1227 09:23:19.447514 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-654859b66d-km642_700034e0-f221-4bca-a973-a6d49c7dc406/barbican-keystone-listener-log/0.log" Dec 27 09:23:19 crc kubenswrapper[4934]: I1227 09:23:19.482989 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-859445d775-g44pr_3f083ef4-9928-400b-969e-1df70b3e9475/barbican-worker/0.log" Dec 27 09:23:19 crc kubenswrapper[4934]: I1227 09:23:19.575728 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-859445d775-g44pr_3f083ef4-9928-400b-969e-1df70b3e9475/barbican-worker-log/0.log" Dec 27 09:23:19 crc kubenswrapper[4934]: I1227 09:23:19.694498 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-8qpb8_68f2fdda-08a1-457f-b61e-c6d68ea69be3/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 27 09:23:19 crc kubenswrapper[4934]: I1227 09:23:19.817177 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3c21d8f2-980c-4491-b91c-a14d5c7eb20f/ceilometer-central-agent/0.log" Dec 27 09:23:19 crc kubenswrapper[4934]: I1227 09:23:19.888763 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3c21d8f2-980c-4491-b91c-a14d5c7eb20f/ceilometer-notification-agent/0.log" Dec 27 09:23:19 crc kubenswrapper[4934]: I1227 09:23:19.902452 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3c21d8f2-980c-4491-b91c-a14d5c7eb20f/proxy-httpd/0.log" Dec 27 09:23:20 crc kubenswrapper[4934]: I1227 09:23:20.005320 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_3c21d8f2-980c-4491-b91c-a14d5c7eb20f/sg-core/0.log" Dec 27 09:23:20 crc kubenswrapper[4934]: I1227 09:23:20.153290 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3981f015-fd8d-4429-b900-ada95f9fffa9/cinder-api-log/0.log" Dec 27 09:23:20 crc kubenswrapper[4934]: I1227 09:23:20.207339 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_3981f015-fd8d-4429-b900-ada95f9fffa9/cinder-api/0.log" Dec 27 09:23:20 crc kubenswrapper[4934]: I1227 09:23:20.350999 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_46b6dbac-6489-4651-8956-8c8a9767a4aa/cinder-scheduler/2.log" Dec 27 09:23:20 crc kubenswrapper[4934]: I1227 09:23:20.359139 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_46b6dbac-6489-4651-8956-8c8a9767a4aa/cinder-scheduler/1.log" Dec 27 09:23:20 crc kubenswrapper[4934]: I1227 09:23:20.430709 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_46b6dbac-6489-4651-8956-8c8a9767a4aa/probe/0.log" Dec 27 09:23:20 crc kubenswrapper[4934]: I1227 09:23:20.542891 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-g2bsq_3d482978-80ac-442d-817f-70f4270892dd/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 27 09:23:20 crc kubenswrapper[4934]: I1227 09:23:20.664967 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-n4z7d_8af50e00-8101-4c06-b3bc-93fd48753773/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 27 09:23:20 crc kubenswrapper[4934]: I1227 09:23:20.722373 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d75f767dc-pv2wg_3e4329ea-695d-49e5-a530-932605451255/init/0.log" Dec 27 09:23:20 crc kubenswrapper[4934]: I1227 09:23:20.966128 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d75f767dc-pv2wg_3e4329ea-695d-49e5-a530-932605451255/init/0.log" Dec 27 09:23:21 crc kubenswrapper[4934]: I1227 09:23:21.026727 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d75f767dc-pv2wg_3e4329ea-695d-49e5-a530-932605451255/dnsmasq-dns/0.log" Dec 27 09:23:21 crc kubenswrapper[4934]: I1227 09:23:21.051943 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-dk7d5_6a50c262-1a91-4dad-91ec-6530d981415a/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 27 09:23:21 crc kubenswrapper[4934]: I1227 09:23:21.254099 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_3152d244-2397-469d-a570-d4fd09ee9a0d/glance-httpd/0.log" Dec 27 09:23:21 crc kubenswrapper[4934]: I1227 09:23:21.286951 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_3152d244-2397-469d-a570-d4fd09ee9a0d/glance-log/0.log" Dec 27 09:23:21 crc kubenswrapper[4934]: I1227 09:23:21.509610 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_3966937c-e242-445a-ac07-f70a6d6a689e/glance-httpd/0.log" Dec 27 09:23:21 crc kubenswrapper[4934]: I1227 09:23:21.523593 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_3966937c-e242-445a-ac07-f70a6d6a689e/glance-log/0.log" Dec 27 09:23:22 crc kubenswrapper[4934]: I1227 09:23:22.224148 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-55c9c9fc75-s5tvt_bb1fd5db-ac54-4e34-9cb3-4caae5f67da8/heat-engine/0.log" Dec 27 09:23:22 crc kubenswrapper[4934]: I1227 09:23:22.374809 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-5wvgb_a5cad0ac-88e1-47e4-8539-da835534e9c5/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 27 09:23:22 crc kubenswrapper[4934]: I1227 09:23:22.422590 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-64f664dd6d-jgmlc_2ccf769c-f8ee-4c19-9ee7-1cd7d839ab43/heat-api/0.log" Dec 27 09:23:22 crc kubenswrapper[4934]: I1227 09:23:22.518719 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-8cb7b99d5-5xbmt_f10e6bef-1760-4a64-a702-675932d1770c/heat-cfnapi/0.log" Dec 27 09:23:22 crc kubenswrapper[4934]: I1227 09:23:22.701069 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-7cw86_41797816-91e0-4a3e-ad30-91da21255b91/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 27 09:23:22 crc kubenswrapper[4934]: I1227 09:23:22.764039 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29447101-6tmzs_81559011-a562-4a12-9f3b-2c6286586406/keystone-cron/0.log" Dec 27 09:23:22 crc kubenswrapper[4934]: I1227 09:23:22.938035 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_4e39152f-3592-4907-9b06-d13074aeb3c4/kube-state-metrics/1.log" Dec 27 09:23:23 crc kubenswrapper[4934]: I1227 09:23:23.037021 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_4e39152f-3592-4907-9b06-d13074aeb3c4/kube-state-metrics/0.log" Dec 27 09:23:23 crc kubenswrapper[4934]: I1227 09:23:23.212210 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-6l2g8_c9000388-2dbb-445d-91ed-06b1952ba4bc/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 27 09:23:23 crc kubenswrapper[4934]: I1227 09:23:23.284983 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_logging-edpm-deployment-openstack-edpm-ipam-mjh9f_4d60d2c6-08ec-4bfc-a066-df42d54ec9f9/logging-edpm-deployment-openstack-edpm-ipam/0.log" Dec 27 09:23:23 crc kubenswrapper[4934]: I1227 09:23:23.533656 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mysqld-exporter-0_bf70522b-7d97-4388-b097-daf31bc4b535/mysqld-exporter/0.log" Dec 27 09:23:23 crc kubenswrapper[4934]: I1227 09:23:23.990915 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-c689587cc-6glpp_72923f32-62d5-49f6-81dc-9741f36139e8/neutron-httpd/0.log" Dec 27 09:23:24 crc kubenswrapper[4934]: I1227 09:23:24.157866 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-c689587cc-6glpp_72923f32-62d5-49f6-81dc-9741f36139e8/neutron-api/0.log" Dec 27 09:23:24 crc kubenswrapper[4934]: I1227 09:23:24.190883 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-255fk_34927e46-8eca-4b2d-ba3a-e68690b804df/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 27 09:23:24 crc kubenswrapper[4934]: I1227 09:23:24.986689 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_4507dd01-8c5e-482f-8881-6e0c70679bec/nova-cell0-conductor-conductor/0.log" Dec 27 09:23:25 crc kubenswrapper[4934]: I1227 09:23:25.054300 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_b102f157-6ec2-4b75-846d-8362a74bf9d2/memcached/0.log" Dec 27 09:23:25 crc kubenswrapper[4934]: I1227 09:23:25.261573 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_22a24b14-793e-404c-9d2e-ac30a2092cab/nova-api-log/0.log" Dec 27 09:23:25 crc kubenswrapper[4934]: I1227 09:23:25.611322 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_22a24b14-793e-404c-9d2e-ac30a2092cab/nova-api-api/0.log" Dec 27 09:23:25 crc kubenswrapper[4934]: I1227 09:23:25.668131 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_33eda6c6-6170-4293-83e0-ecce87ddca8c/nova-cell1-conductor-conductor/0.log" Dec 27 09:23:25 crc kubenswrapper[4934]: I1227 09:23:25.798518 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_be9c97e5-5a63-44cd-854f-16a1bcbbeb35/nova-cell1-novncproxy-novncproxy/0.log" Dec 27 09:23:25 crc kubenswrapper[4934]: I1227 09:23:25.978156 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-qb7m2_2f4288f0-95f7-4b96-8066-e8b892d401ae/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 27 09:23:26 crc kubenswrapper[4934]: I1227 09:23:26.174131 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8094d1d3-d139-4889-84aa-2ba422a145db/nova-metadata-log/0.log" Dec 27 09:23:26 crc kubenswrapper[4934]: I1227 09:23:26.517704 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_07762cba-4cfa-43ec-b661-bf478a62d317/nova-scheduler-scheduler/0.log" Dec 27 09:23:26 crc kubenswrapper[4934]: I1227 09:23:26.694788 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ca95a4ef-282d-4c26-983f-d7d00f90443c/mysql-bootstrap/0.log" Dec 27 09:23:27 crc kubenswrapper[4934]: I1227 09:23:27.766383 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ca95a4ef-282d-4c26-983f-d7d00f90443c/galera/0.log" Dec 27 09:23:27 crc kubenswrapper[4934]: I1227 09:23:27.820693 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ca95a4ef-282d-4c26-983f-d7d00f90443c/mysql-bootstrap/0.log" Dec 27 09:23:27 crc kubenswrapper[4934]: I1227 09:23:27.828074 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ca95a4ef-282d-4c26-983f-d7d00f90443c/galera/1.log" Dec 27 09:23:28 crc kubenswrapper[4934]: I1227 09:23:28.123514 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d03cf582-2192-4a10-b21e-10befebde419/mysql-bootstrap/0.log" Dec 27 09:23:28 crc kubenswrapper[4934]: I1227 09:23:28.356904 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d03cf582-2192-4a10-b21e-10befebde419/mysql-bootstrap/0.log" Dec 27 09:23:28 crc kubenswrapper[4934]: I1227 09:23:28.366943 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d03cf582-2192-4a10-b21e-10befebde419/galera/0.log" Dec 27 09:23:28 crc kubenswrapper[4934]: I1227 09:23:28.425957 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_d03cf582-2192-4a10-b21e-10befebde419/galera/1.log" Dec 27 09:23:28 crc kubenswrapper[4934]: I1227 09:23:28.574036 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8094d1d3-d139-4889-84aa-2ba422a145db/nova-metadata-metadata/0.log" Dec 27 09:23:28 crc kubenswrapper[4934]: I1227 09:23:28.614980 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_f5eabd4f-a103-4ec0-9c14-7deb0e36fe61/openstackclient/0.log" Dec 27 09:23:28 crc kubenswrapper[4934]: I1227 09:23:28.781467 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-22dwz_f4b9eef2-987b-4c54-803c-6655aceab8f6/ovn-controller/0.log" Dec 27 09:23:28 crc kubenswrapper[4934]: I1227 09:23:28.849528 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-5v4n5_36e93e22-f392-4011-a82a-a5d1c26b5970/openstack-network-exporter/0.log" Dec 27 09:23:29 crc kubenswrapper[4934]: I1227 09:23:29.063162 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bdjjv_54ab7f4e-f623-473b-8b6a-3363fe487b4e/ovsdb-server-init/0.log" Dec 27 09:23:29 crc kubenswrapper[4934]: I1227 09:23:29.259872 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bdjjv_54ab7f4e-f623-473b-8b6a-3363fe487b4e/ovsdb-server-init/0.log" Dec 27 09:23:29 crc kubenswrapper[4934]: I1227 09:23:29.262876 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bdjjv_54ab7f4e-f623-473b-8b6a-3363fe487b4e/ovs-vswitchd/0.log" Dec 27 09:23:29 crc kubenswrapper[4934]: I1227 09:23:29.290597 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bdjjv_54ab7f4e-f623-473b-8b6a-3363fe487b4e/ovsdb-server/0.log" Dec 27 09:23:29 crc kubenswrapper[4934]: I1227 09:23:29.526808 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_727f882e-fa63-46e9-9cd5-9781e472f2ef/openstack-network-exporter/0.log" Dec 27 09:23:29 crc kubenswrapper[4934]: I1227 09:23:29.573636 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-fcwdp_595c2188-f2bb-4c19-b1a9-2da10bdc89c7/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 27 09:23:29 crc kubenswrapper[4934]: I1227 09:23:29.576373 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_727f882e-fa63-46e9-9cd5-9781e472f2ef/ovn-northd/0.log" Dec 27 09:23:29 crc kubenswrapper[4934]: I1227 09:23:29.769956 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_cc83d812-aa8d-45e4-b6df-02e6f8a9d9db/openstack-network-exporter/0.log" Dec 27 09:23:29 crc kubenswrapper[4934]: I1227 09:23:29.844402 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_cc83d812-aa8d-45e4-b6df-02e6f8a9d9db/ovsdbserver-nb/0.log" Dec 27 09:23:29 crc kubenswrapper[4934]: I1227 09:23:29.986194 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_049be8b3-9033-425c-940e-4a36854f06d9/openstack-network-exporter/0.log" Dec 27 09:23:30 crc kubenswrapper[4934]: I1227 09:23:30.035302 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_049be8b3-9033-425c-940e-4a36854f06d9/ovsdbserver-sb/0.log" Dec 27 09:23:30 crc kubenswrapper[4934]: I1227 09:23:30.234857 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5c6bd9586d-nf2kk_5259d7a2-5596-4cbf-88f1-2b0d59cf2b74/placement-api/0.log" Dec 27 09:23:30 crc kubenswrapper[4934]: I1227 09:23:30.388854 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5c6bd9586d-nf2kk_5259d7a2-5596-4cbf-88f1-2b0d59cf2b74/placement-log/0.log" Dec 27 09:23:30 crc kubenswrapper[4934]: I1227 09:23:30.425007 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_e26e3510-3438-4cb9-8d00-f2d0fdf33ef5/init-config-reloader/0.log" Dec 27 09:23:30 crc kubenswrapper[4934]: I1227 09:23:30.530297 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7fd8c76fd8-kbl58_1de4b3bd-106e-44ce-b68e-8a2c28916b9b/keystone-api/0.log" Dec 27 09:23:30 crc kubenswrapper[4934]: I1227 09:23:30.622305 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_e26e3510-3438-4cb9-8d00-f2d0fdf33ef5/init-config-reloader/0.log" Dec 27 09:23:30 crc kubenswrapper[4934]: I1227 09:23:30.644808 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_e26e3510-3438-4cb9-8d00-f2d0fdf33ef5/thanos-sidecar/0.log" Dec 27 09:23:30 crc kubenswrapper[4934]: I1227 09:23:30.644905 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_e26e3510-3438-4cb9-8d00-f2d0fdf33ef5/config-reloader/0.log" Dec 27 09:23:30 crc kubenswrapper[4934]: I1227 09:23:30.647188 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_e26e3510-3438-4cb9-8d00-f2d0fdf33ef5/prometheus/0.log" Dec 27 09:23:30 crc kubenswrapper[4934]: I1227 09:23:30.797177 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_85c58dd9-a367-4468-ad44-0f567a89c903/setup-container/0.log" Dec 27 09:23:30 crc kubenswrapper[4934]: I1227 09:23:30.964691 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_85c58dd9-a367-4468-ad44-0f567a89c903/setup-container/0.log" Dec 27 09:23:31 crc kubenswrapper[4934]: I1227 09:23:31.001053 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_85c58dd9-a367-4468-ad44-0f567a89c903/rabbitmq/0.log" Dec 27 09:23:31 crc kubenswrapper[4934]: I1227 09:23:31.015819 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b6a52bce-47cf-48e4-9f9a-7b3d978dee58/setup-container/0.log" Dec 27 09:23:31 crc kubenswrapper[4934]: I1227 09:23:31.214111 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b6a52bce-47cf-48e4-9f9a-7b3d978dee58/setup-container/0.log" Dec 27 09:23:31 crc kubenswrapper[4934]: I1227 09:23:31.266173 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-1_93d5ae97-eb77-4368-931c-0421b65ac057/setup-container/0.log" Dec 27 09:23:31 crc kubenswrapper[4934]: I1227 09:23:31.274268 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b6a52bce-47cf-48e4-9f9a-7b3d978dee58/rabbitmq/0.log" Dec 27 09:23:31 crc kubenswrapper[4934]: I1227 09:23:31.500176 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-1_93d5ae97-eb77-4368-931c-0421b65ac057/setup-container/0.log" Dec 27 09:23:31 crc kubenswrapper[4934]: I1227 09:23:31.535648 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-1_93d5ae97-eb77-4368-931c-0421b65ac057/rabbitmq/0.log" Dec 27 09:23:31 crc kubenswrapper[4934]: I1227 09:23:31.568178 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-2_5a3ecad3-f6df-4051-be94-faff5cbc4e1f/setup-container/0.log" Dec 27 09:23:31 crc kubenswrapper[4934]: I1227 09:23:31.730070 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-2_5a3ecad3-f6df-4051-be94-faff5cbc4e1f/rabbitmq/0.log" Dec 27 09:23:31 crc kubenswrapper[4934]: I1227 09:23:31.760122 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-2_5a3ecad3-f6df-4051-be94-faff5cbc4e1f/setup-container/0.log" Dec 27 09:23:31 crc kubenswrapper[4934]: I1227 09:23:31.794690 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-l6kg8_a05cb14c-f210-4e43-8a1a-95fb4e86d55b/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 27 09:23:31 crc kubenswrapper[4934]: I1227 09:23:31.911341 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-mrdpt_61f9c149-e7a9-4ca5-b708-aa9274ee74af/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 27 09:23:31 crc kubenswrapper[4934]: I1227 09:23:31.980931 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-lsg9b_3833a830-a283-4755-8909-c5a5c7a218a9/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 27 09:23:32 crc kubenswrapper[4934]: I1227 09:23:32.057359 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-j9cms_19b8530e-a7de-43d2-bfa7-b0f41ff55156/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 27 09:23:32 crc kubenswrapper[4934]: I1227 09:23:32.184907 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-d6vnn_66f3fb1f-e2f4-40b4-bdf7-c67abdf2cc35/ssh-known-hosts-edpm-deployment/0.log" Dec 27 09:23:32 crc kubenswrapper[4934]: I1227 09:23:32.314571 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-788c8bbcbc-rtffl_930bd1af-796b-409e-8bae-d828b871edad/proxy-server/0.log" Dec 27 09:23:32 crc kubenswrapper[4934]: I1227 09:23:32.431857 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-788c8bbcbc-rtffl_930bd1af-796b-409e-8bae-d828b871edad/proxy-httpd/0.log" Dec 27 09:23:32 crc kubenswrapper[4934]: I1227 09:23:32.441855 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-bswkq_ccb12f0b-32d5-4ab4-9228-a8aa005b1304/swift-ring-rebalance/0.log" Dec 27 09:23:32 crc kubenswrapper[4934]: I1227 09:23:32.545974 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e59489eb-aaa1-4188-b593-d3d3c2f1f412/account-auditor/0.log" Dec 27 09:23:32 crc kubenswrapper[4934]: I1227 09:23:32.635660 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e59489eb-aaa1-4188-b593-d3d3c2f1f412/account-reaper/0.log" Dec 27 09:23:32 crc kubenswrapper[4934]: I1227 09:23:32.679732 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e59489eb-aaa1-4188-b593-d3d3c2f1f412/account-server/0.log" Dec 27 09:23:32 crc kubenswrapper[4934]: I1227 09:23:32.693026 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e59489eb-aaa1-4188-b593-d3d3c2f1f412/account-replicator/0.log" Dec 27 09:23:32 crc kubenswrapper[4934]: I1227 09:23:32.749171 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e59489eb-aaa1-4188-b593-d3d3c2f1f412/container-auditor/0.log" Dec 27 09:23:32 crc kubenswrapper[4934]: I1227 09:23:32.831258 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e59489eb-aaa1-4188-b593-d3d3c2f1f412/container-replicator/0.log" Dec 27 09:23:32 crc kubenswrapper[4934]: I1227 09:23:32.849674 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e59489eb-aaa1-4188-b593-d3d3c2f1f412/container-server/0.log" Dec 27 09:23:32 crc kubenswrapper[4934]: I1227 09:23:32.891009 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e59489eb-aaa1-4188-b593-d3d3c2f1f412/container-updater/0.log" Dec 27 09:23:32 crc kubenswrapper[4934]: I1227 09:23:32.922162 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e59489eb-aaa1-4188-b593-d3d3c2f1f412/object-auditor/0.log" Dec 27 09:23:32 crc kubenswrapper[4934]: I1227 09:23:32.983912 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e59489eb-aaa1-4188-b593-d3d3c2f1f412/object-expirer/0.log" Dec 27 09:23:33 crc kubenswrapper[4934]: I1227 09:23:33.041433 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e59489eb-aaa1-4188-b593-d3d3c2f1f412/object-server/0.log" Dec 27 09:23:33 crc kubenswrapper[4934]: I1227 09:23:33.058766 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e59489eb-aaa1-4188-b593-d3d3c2f1f412/object-replicator/0.log" Dec 27 09:23:33 crc kubenswrapper[4934]: I1227 09:23:33.111547 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e59489eb-aaa1-4188-b593-d3d3c2f1f412/object-updater/0.log" Dec 27 09:23:33 crc kubenswrapper[4934]: I1227 09:23:33.170268 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e59489eb-aaa1-4188-b593-d3d3c2f1f412/rsync/0.log" Dec 27 09:23:33 crc kubenswrapper[4934]: I1227 09:23:33.216943 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e59489eb-aaa1-4188-b593-d3d3c2f1f412/swift-recon-cron/0.log" Dec 27 09:23:33 crc kubenswrapper[4934]: I1227 09:23:33.324971 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-qrqkm_b87a1e52-c2d2-4384-97ee-c3cd68e90386/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 27 09:23:33 crc kubenswrapper[4934]: I1227 09:23:33.390751 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-power-monitoring-edpm-deployment-openstack-edpm-758dw_54b601bd-dc4a-49b3-bd37-a6d13ae07cdf/telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam/0.log" Dec 27 09:23:33 crc kubenswrapper[4934]: I1227 09:23:33.559726 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_aa0bb6c7-0651-4833-b1e0-364be9987257/test-operator-logs-container/0.log" Dec 27 09:23:33 crc kubenswrapper[4934]: I1227 09:23:33.703133 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-s8gxz_6174eb59-8d05-409e-a838-807ef965ee5c/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 27 09:23:34 crc kubenswrapper[4934]: I1227 09:23:34.105297 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_d38686ee-0d94-4694-845a-e1ee5971af6f/tempest-tests-tempest-tests-runner/0.log" Dec 27 09:24:03 crc kubenswrapper[4934]: I1227 09:24:03.257110 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-649cf68dc9-t4bqz_dc4ddef4-e9a5-4934-846e-1d1bcfe9f223/kube-rbac-proxy/0.log" Dec 27 09:24:03 crc kubenswrapper[4934]: I1227 09:24:03.346804 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-649cf68dc9-t4bqz_dc4ddef4-e9a5-4934-846e-1d1bcfe9f223/manager/0.log" Dec 27 09:24:03 crc kubenswrapper[4934]: I1227 09:24:03.350368 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-649cf68dc9-t4bqz_dc4ddef4-e9a5-4934-846e-1d1bcfe9f223/manager/1.log" Dec 27 09:24:03 crc kubenswrapper[4934]: I1227 09:24:03.487610 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-669b58f65-t89p4_935a4c26-2678-4ead-87ee-1297850d89ff/kube-rbac-proxy/0.log" Dec 27 09:24:03 crc kubenswrapper[4934]: I1227 09:24:03.619655 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-669b58f65-t89p4_935a4c26-2678-4ead-87ee-1297850d89ff/manager/0.log" Dec 27 09:24:03 crc kubenswrapper[4934]: I1227 09:24:03.711661 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-69977bdf55-p5jgx_fc1c7052-d894-4fa1-b6bb-951cad9e32e0/kube-rbac-proxy/0.log" Dec 27 09:24:03 crc kubenswrapper[4934]: I1227 09:24:03.757173 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-69977bdf55-p5jgx_fc1c7052-d894-4fa1-b6bb-951cad9e32e0/manager/0.log" Dec 27 09:24:03 crc kubenswrapper[4934]: I1227 09:24:03.845958 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv_3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4/util/0.log" Dec 27 09:24:04 crc kubenswrapper[4934]: I1227 09:24:04.034001 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv_3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4/pull/0.log" Dec 27 09:24:04 crc kubenswrapper[4934]: I1227 09:24:04.047151 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv_3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4/util/0.log" Dec 27 09:24:04 crc kubenswrapper[4934]: I1227 09:24:04.063549 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv_3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4/pull/0.log" Dec 27 09:24:04 crc kubenswrapper[4934]: I1227 09:24:04.269360 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv_3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4/extract/0.log" Dec 27 09:24:04 crc kubenswrapper[4934]: I1227 09:24:04.278325 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv_3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4/pull/0.log" Dec 27 09:24:04 crc kubenswrapper[4934]: I1227 09:24:04.288949 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea168dbd2ded9be20f22262dd3a5021ac4f7db043021fb2084bb8bda0a9nhbv_3fb5ccd1-1f57-4b3b-97d2-97af95b73ec4/util/0.log" Dec 27 09:24:04 crc kubenswrapper[4934]: I1227 09:24:04.491759 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-64fb555449-tbmpv_3df60bce-0768-4949-8b8d-3bb28566cd4a/kube-rbac-proxy/0.log" Dec 27 09:24:04 crc kubenswrapper[4934]: I1227 09:24:04.565592 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-65cdd6cc59-ggc7k_77029090-da44-403b-a20f-d72105281956/kube-rbac-proxy/0.log" Dec 27 09:24:04 crc kubenswrapper[4934]: I1227 09:24:04.573614 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-64fb555449-tbmpv_3df60bce-0768-4949-8b8d-3bb28566cd4a/manager/0.log" Dec 27 09:24:04 crc kubenswrapper[4934]: I1227 09:24:04.706941 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-65cdd6cc59-ggc7k_77029090-da44-403b-a20f-d72105281956/manager/1.log" Dec 27 09:24:04 crc kubenswrapper[4934]: I1227 09:24:04.800952 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-65cdd6cc59-ggc7k_77029090-da44-403b-a20f-d72105281956/manager/0.log" Dec 27 09:24:04 crc kubenswrapper[4934]: I1227 09:24:04.822340 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54996ddd58-rjkpr_9fbceb4d-e7ae-42c0-8a3f-17807bc9f2ca/kube-rbac-proxy/0.log" Dec 27 09:24:04 crc kubenswrapper[4934]: I1227 09:24:04.898103 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54996ddd58-rjkpr_9fbceb4d-e7ae-42c0-8a3f-17807bc9f2ca/manager/1.log" Dec 27 09:24:04 crc kubenswrapper[4934]: I1227 09:24:04.982029 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-54996ddd58-rjkpr_9fbceb4d-e7ae-42c0-8a3f-17807bc9f2ca/manager/0.log" Dec 27 09:24:05 crc kubenswrapper[4934]: I1227 09:24:05.016204 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6c77d5f585-lg74h_e2db190a-b455-44fc-a43e-2677e5af27b2/kube-rbac-proxy/0.log" Dec 27 09:24:05 crc kubenswrapper[4934]: I1227 09:24:05.120832 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6c77d5f585-lg74h_e2db190a-b455-44fc-a43e-2677e5af27b2/manager/1.log" Dec 27 09:24:05 crc kubenswrapper[4934]: I1227 09:24:05.266554 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-7564bdd644-bg78w_f41746ba-5695-41bc-8078-29ba3d80e3f1/kube-rbac-proxy/0.log" Dec 27 09:24:05 crc kubenswrapper[4934]: I1227 09:24:05.370822 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-6c77d5f585-lg74h_e2db190a-b455-44fc-a43e-2677e5af27b2/manager/0.log" Dec 27 09:24:05 crc kubenswrapper[4934]: I1227 09:24:05.375119 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-7564bdd644-bg78w_f41746ba-5695-41bc-8078-29ba3d80e3f1/manager/1.log" Dec 27 09:24:05 crc kubenswrapper[4934]: I1227 09:24:05.489004 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-7564bdd644-bg78w_f41746ba-5695-41bc-8078-29ba3d80e3f1/manager/0.log" Dec 27 09:24:05 crc kubenswrapper[4934]: I1227 09:24:05.565422 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7f764db9b-pgjkp_718122b6-c9f4-4d6f-92b5-6862c15f4ee6/kube-rbac-proxy/0.log" Dec 27 09:24:05 crc kubenswrapper[4934]: I1227 09:24:05.657559 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7f764db9b-pgjkp_718122b6-c9f4-4d6f-92b5-6862c15f4ee6/manager/1.log" Dec 27 09:24:05 crc kubenswrapper[4934]: I1227 09:24:05.778206 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7cc599445b-dlwjl_37d56061-eab6-43e5-8dec-f59220da5d0d/kube-rbac-proxy/0.log" Dec 27 09:24:05 crc kubenswrapper[4934]: I1227 09:24:05.837931 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7f764db9b-pgjkp_718122b6-c9f4-4d6f-92b5-6862c15f4ee6/manager/0.log" Dec 27 09:24:05 crc kubenswrapper[4934]: I1227 09:24:05.859851 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7cc599445b-dlwjl_37d56061-eab6-43e5-8dec-f59220da5d0d/manager/1.log" Dec 27 09:24:05 crc kubenswrapper[4934]: I1227 09:24:05.929395 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7cc599445b-dlwjl_37d56061-eab6-43e5-8dec-f59220da5d0d/manager/0.log" Dec 27 09:24:06 crc kubenswrapper[4934]: I1227 09:24:06.033690 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-64d7c556cd-ffx59_828ebb68-326b-4c32-bc0a-d7f258b45ebd/kube-rbac-proxy/0.log" Dec 27 09:24:06 crc kubenswrapper[4934]: I1227 09:24:06.090825 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-64d7c556cd-ffx59_828ebb68-326b-4c32-bc0a-d7f258b45ebd/manager/0.log" Dec 27 09:24:06 crc kubenswrapper[4934]: I1227 09:24:06.185162 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-58879495c-d27qh_89ea143d-f68d-46fc-b15a-6e00a418f65c/kube-rbac-proxy/0.log" Dec 27 09:24:06 crc kubenswrapper[4934]: I1227 09:24:06.278544 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-58879495c-d27qh_89ea143d-f68d-46fc-b15a-6e00a418f65c/manager/0.log" Dec 27 09:24:06 crc kubenswrapper[4934]: I1227 09:24:06.353456 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6b444986fd-djrtj_d090706d-34a7-4c72-9cb9-ad0601db35a6/kube-rbac-proxy/0.log" Dec 27 09:24:06 crc kubenswrapper[4934]: I1227 09:24:06.431232 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6b444986fd-djrtj_d090706d-34a7-4c72-9cb9-ad0601db35a6/manager/1.log" Dec 27 09:24:06 crc kubenswrapper[4934]: I1227 09:24:06.557163 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6b444986fd-djrtj_d090706d-34a7-4c72-9cb9-ad0601db35a6/manager/0.log" Dec 27 09:24:06 crc kubenswrapper[4934]: I1227 09:24:06.591249 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-d5fb87cb8-w2xgr_1c0960f0-ef55-479a-b375-a09e69254743/kube-rbac-proxy/0.log" Dec 27 09:24:06 crc kubenswrapper[4934]: I1227 09:24:06.653105 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-d5fb87cb8-w2xgr_1c0960f0-ef55-479a-b375-a09e69254743/manager/1.log" Dec 27 09:24:06 crc kubenswrapper[4934]: I1227 09:24:06.698488 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-d5fb87cb8-w2xgr_1c0960f0-ef55-479a-b375-a09e69254743/manager/0.log" Dec 27 09:24:06 crc kubenswrapper[4934]: I1227 09:24:06.792702 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc_b5ea792c-b43f-4221-b7bb-aba3421de3d8/kube-rbac-proxy/0.log" Dec 27 09:24:06 crc kubenswrapper[4934]: I1227 09:24:06.812135 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc_b5ea792c-b43f-4221-b7bb-aba3421de3d8/manager/1.log" Dec 27 09:24:06 crc kubenswrapper[4934]: I1227 09:24:06.881259 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cbb4f75bdwf5zc_b5ea792c-b43f-4221-b7bb-aba3421de3d8/manager/0.log" Dec 27 09:24:06 crc kubenswrapper[4934]: I1227 09:24:06.961303 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5758858f4d-mg9wf_e44e23aa-f466-41cb-b855-d0bdc25cf05b/kube-rbac-proxy/0.log" Dec 27 09:24:07 crc kubenswrapper[4934]: I1227 09:24:07.141239 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-78479c6bd7-j8dhc_b461380d-6573-4898-9bc8-c82ceba822d8/kube-rbac-proxy/0.log" Dec 27 09:24:07 crc kubenswrapper[4934]: I1227 09:24:07.250825 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-v7q5s_a41afe73-59e9-4cbe-b2e7-1f755767daf5/registry-server/0.log" Dec 27 09:24:07 crc kubenswrapper[4934]: I1227 09:24:07.392117 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-78479c6bd7-j8dhc_b461380d-6573-4898-9bc8-c82ceba822d8/operator/0.log" Dec 27 09:24:07 crc kubenswrapper[4934]: I1227 09:24:07.475014 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5b67cfc8fb-6r7n4_59d65d09-c15c-49d9-8ee7-472ffe0dda55/kube-rbac-proxy/0.log" Dec 27 09:24:07 crc kubenswrapper[4934]: I1227 09:24:07.501546 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5b67cfc8fb-6r7n4_59d65d09-c15c-49d9-8ee7-472ffe0dda55/manager/1.log" Dec 27 09:24:07 crc kubenswrapper[4934]: I1227 09:24:07.681144 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5b67cfc8fb-6r7n4_59d65d09-c15c-49d9-8ee7-472ffe0dda55/manager/0.log" Dec 27 09:24:07 crc kubenswrapper[4934]: I1227 09:24:07.747591 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-cc776f956-cj2c7_7d49c3c9-83c2-4c42-a318-c9d41ddf1bf8/kube-rbac-proxy/0.log" Dec 27 09:24:07 crc kubenswrapper[4934]: I1227 09:24:07.751675 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-cc776f956-cj2c7_7d49c3c9-83c2-4c42-a318-c9d41ddf1bf8/manager/0.log" Dec 27 09:24:08 crc kubenswrapper[4934]: I1227 09:24:08.017068 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-7c9ff8845d-ghvpg_4f825c5b-90ef-4b56-a759-d7ccbb312bed/kube-rbac-proxy/0.log" Dec 27 09:24:08 crc kubenswrapper[4934]: I1227 09:24:08.032852 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-z9x75_e5e99871-c4b8-4bde-b7b7-41b7b5e2f5b0/operator/0.log" Dec 27 09:24:08 crc kubenswrapper[4934]: I1227 09:24:08.196681 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-7c9ff8845d-ghvpg_4f825c5b-90ef-4b56-a759-d7ccbb312bed/manager/0.log" Dec 27 09:24:08 crc kubenswrapper[4934]: I1227 09:24:08.250816 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6659cb75fc-9kzpj_f2d5e345-1bd9-4313-bdae-7637cb7ed944/kube-rbac-proxy/0.log" Dec 27 09:24:08 crc kubenswrapper[4934]: I1227 09:24:08.295127 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6659cb75fc-9kzpj_f2d5e345-1bd9-4313-bdae-7637cb7ed944/manager/1.log" Dec 27 09:24:08 crc kubenswrapper[4934]: I1227 09:24:08.450841 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5d79c6465c-zldvm_f1cce183-e0a8-4cf1-854e-7a60986f194f/kube-rbac-proxy/0.log" Dec 27 09:24:08 crc kubenswrapper[4934]: I1227 09:24:08.503726 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5758858f4d-mg9wf_e44e23aa-f466-41cb-b855-d0bdc25cf05b/manager/0.log" Dec 27 09:24:08 crc kubenswrapper[4934]: I1227 09:24:08.531564 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5d79c6465c-zldvm_f1cce183-e0a8-4cf1-854e-7a60986f194f/manager/1.log" Dec 27 09:24:08 crc kubenswrapper[4934]: I1227 09:24:08.729616 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-59cf64b5bd-28lln_35d6eb8d-3797-432c-a5ea-65f31adf7262/kube-rbac-proxy/0.log" Dec 27 09:24:08 crc kubenswrapper[4934]: I1227 09:24:08.744886 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5d79c6465c-zldvm_f1cce183-e0a8-4cf1-854e-7a60986f194f/manager/0.log" Dec 27 09:24:08 crc kubenswrapper[4934]: I1227 09:24:08.756195 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6659cb75fc-9kzpj_f2d5e345-1bd9-4313-bdae-7637cb7ed944/manager/0.log" Dec 27 09:24:08 crc kubenswrapper[4934]: I1227 09:24:08.802298 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-59cf64b5bd-28lln_35d6eb8d-3797-432c-a5ea-65f31adf7262/manager/1.log" Dec 27 09:24:08 crc kubenswrapper[4934]: I1227 09:24:08.920320 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-59cf64b5bd-28lln_35d6eb8d-3797-432c-a5ea-65f31adf7262/manager/0.log" Dec 27 09:24:28 crc kubenswrapper[4934]: I1227 09:24:28.288056 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zchqn_e4f56101-c810-4e1f-96e8-c7cbcb2ec719/control-plane-machine-set-operator/0.log" Dec 27 09:24:28 crc kubenswrapper[4934]: I1227 09:24:28.516383 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-whhqx_46527d26-eb38-494f-85f7-f5f56b8dfada/machine-api-operator/0.log" Dec 27 09:24:28 crc kubenswrapper[4934]: I1227 09:24:28.530851 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-whhqx_46527d26-eb38-494f-85f7-f5f56b8dfada/kube-rbac-proxy/0.log" Dec 27 09:24:31 crc kubenswrapper[4934]: I1227 09:24:31.356700 4934 scope.go:117] "RemoveContainer" containerID="b0482879005ff56387cb38df22c6a3856d9b92fc44414f4ba7c4df70aa7d5a2c" Dec 27 09:24:31 crc kubenswrapper[4934]: I1227 09:24:31.386804 4934 scope.go:117] "RemoveContainer" containerID="12764a36c08b67337808b28e847cc266ca1713e84de57f838504d5c559993b85" Dec 27 09:24:31 crc kubenswrapper[4934]: I1227 09:24:31.450128 4934 scope.go:117] "RemoveContainer" containerID="9fb9a9a105f82b1c3ef2cdc89348c9929df9be11c73f572ecd9fa03130a17376" Dec 27 09:24:42 crc kubenswrapper[4934]: I1227 09:24:42.749960 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-26fts_1c38c7d2-7994-4d2b-83a2-d98066e4ba5b/cert-manager-controller/0.log" Dec 27 09:24:42 crc kubenswrapper[4934]: I1227 09:24:42.941547 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-hmzlk_e14d5680-64cf-4640-a505-62020b9c31f4/cert-manager-cainjector/0.log" Dec 27 09:24:42 crc kubenswrapper[4934]: I1227 09:24:42.977811 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-5kwg4_dab3e53b-cbb2-4423-a8e3-432972ecedc3/cert-manager-webhook/0.log" Dec 27 09:24:57 crc kubenswrapper[4934]: I1227 09:24:57.054095 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6ff7998486-7hbj6_60914d81-fa54-4836-8b55-2cbe909df91d/nmstate-console-plugin/0.log" Dec 27 09:24:57 crc kubenswrapper[4934]: I1227 09:24:57.234118 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-kbxfg_224f9a57-4665-4c0f-865c-ac2267f69b7e/nmstate-handler/0.log" Dec 27 09:24:57 crc kubenswrapper[4934]: I1227 09:24:57.321012 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f7f7578db-6fp5z_4f0bb073-052f-45eb-9cd6-020b9323b2ca/kube-rbac-proxy/0.log" Dec 27 09:24:57 crc kubenswrapper[4934]: I1227 09:24:57.327979 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f7f7578db-6fp5z_4f0bb073-052f-45eb-9cd6-020b9323b2ca/nmstate-metrics/0.log" Dec 27 09:24:57 crc kubenswrapper[4934]: I1227 09:24:57.463962 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-6769fb99d-p29kj_e7bfbaac-ef25-4807-8337-14656ff53f15/nmstate-operator/0.log" Dec 27 09:24:57 crc kubenswrapper[4934]: I1227 09:24:57.562670 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-f8fb84555-fwjdt_415e9aa6-5a7a-4d93-924d-0213c9a5ca4d/nmstate-webhook/0.log" Dec 27 09:25:11 crc kubenswrapper[4934]: I1227 09:25:11.428400 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-77c7786d5-gz9d9_00799e47-9b94-49bf-8bd1-d6bb5036285a/kube-rbac-proxy/0.log" Dec 27 09:25:11 crc kubenswrapper[4934]: I1227 09:25:11.509839 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-77c7786d5-gz9d9_00799e47-9b94-49bf-8bd1-d6bb5036285a/manager/1.log" Dec 27 09:25:11 crc kubenswrapper[4934]: I1227 09:25:11.621837 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-77c7786d5-gz9d9_00799e47-9b94-49bf-8bd1-d6bb5036285a/manager/0.log" Dec 27 09:25:15 crc kubenswrapper[4934]: I1227 09:25:15.330155 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 09:25:15 crc kubenswrapper[4934]: I1227 09:25:15.330807 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 09:25:26 crc kubenswrapper[4934]: I1227 09:25:26.399286 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-79cf69ddc8-h47cv_8b6af0c6-05e2-4adc-8e98-7d2e4f6254ea/cluster-logging-operator/0.log" Dec 27 09:25:26 crc kubenswrapper[4934]: I1227 09:25:26.563431 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-dngrf_dc561c93-57e5-4231-9f5d-6b93101c4b38/collector/0.log" Dec 27 09:25:26 crc kubenswrapper[4934]: I1227 09:25:26.640907 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_e6de9198-9266-43e8-b760-e21f0afa885e/loki-compactor/0.log" Dec 27 09:25:26 crc kubenswrapper[4934]: I1227 09:25:26.769555 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-5f678c8dd6-nqqn7_5864b138-b8ef-405e-9ed9-be39f8e13e8d/loki-distributor/0.log" Dec 27 09:25:26 crc kubenswrapper[4934]: I1227 09:25:26.866640 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-74dbf44c45-4ll7v_eac5b26c-d598-4c68-95b5-e2583b456af9/gateway/0.log" Dec 27 09:25:26 crc kubenswrapper[4934]: I1227 09:25:26.877189 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-74dbf44c45-4ll7v_eac5b26c-d598-4c68-95b5-e2583b456af9/opa/0.log" Dec 27 09:25:27 crc kubenswrapper[4934]: I1227 09:25:27.037543 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-74dbf44c45-qd69g_47cbee2e-c325-4a8d-a26d-94bced60ee78/gateway/0.log" Dec 27 09:25:27 crc kubenswrapper[4934]: I1227 09:25:27.072506 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-74dbf44c45-qd69g_47cbee2e-c325-4a8d-a26d-94bced60ee78/opa/0.log" Dec 27 09:25:27 crc kubenswrapper[4934]: I1227 09:25:27.184593 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_01f91788-9f82-4508-906b-1e98c9e05c2c/loki-index-gateway/0.log" Dec 27 09:25:27 crc kubenswrapper[4934]: I1227 09:25:27.334287 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_642f3213-8063-4ade-800e-4e05ba16d642/loki-ingester/0.log" Dec 27 09:25:27 crc kubenswrapper[4934]: I1227 09:25:27.387711 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-76788598db-6wrwh_ae25e4b9-7a2a-41ff-9944-d6c45603cbec/loki-querier/0.log" Dec 27 09:25:27 crc kubenswrapper[4934]: I1227 09:25:27.540022 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-69d9546745-4b9rt_b9e3cc00-7197-4c75-a732-136e65d893c8/loki-query-frontend/0.log" Dec 27 09:25:41 crc kubenswrapper[4934]: I1227 09:25:41.632620 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-7c5ls_a8825499-558f-4b31-b09b-0dcc92662f65/controller/1.log" Dec 27 09:25:41 crc kubenswrapper[4934]: I1227 09:25:41.694870 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-7c5ls_a8825499-558f-4b31-b09b-0dcc92662f65/controller/0.log" Dec 27 09:25:41 crc kubenswrapper[4934]: I1227 09:25:41.936679 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-5bddd4b946-7c5ls_a8825499-558f-4b31-b09b-0dcc92662f65/kube-rbac-proxy/0.log" Dec 27 09:25:41 crc kubenswrapper[4934]: I1227 09:25:41.957058 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2lll_c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd/cp-frr-files/0.log" Dec 27 09:25:42 crc kubenswrapper[4934]: I1227 09:25:42.164010 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2lll_c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd/cp-frr-files/0.log" Dec 27 09:25:42 crc kubenswrapper[4934]: I1227 09:25:42.179722 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2lll_c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd/cp-reloader/0.log" Dec 27 09:25:42 crc kubenswrapper[4934]: I1227 09:25:42.195392 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2lll_c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd/cp-reloader/0.log" Dec 27 09:25:42 crc kubenswrapper[4934]: I1227 09:25:42.209144 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2lll_c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd/cp-metrics/0.log" Dec 27 09:25:42 crc kubenswrapper[4934]: I1227 09:25:42.386628 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2lll_c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd/cp-reloader/0.log" Dec 27 09:25:42 crc kubenswrapper[4934]: I1227 09:25:42.394523 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2lll_c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd/cp-metrics/0.log" Dec 27 09:25:42 crc kubenswrapper[4934]: I1227 09:25:42.401222 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2lll_c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd/cp-frr-files/0.log" Dec 27 09:25:42 crc kubenswrapper[4934]: I1227 09:25:42.410460 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2lll_c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd/cp-metrics/0.log" Dec 27 09:25:42 crc kubenswrapper[4934]: I1227 09:25:42.574162 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2lll_c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd/cp-frr-files/0.log" Dec 27 09:25:42 crc kubenswrapper[4934]: I1227 09:25:42.606770 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2lll_c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd/cp-reloader/0.log" Dec 27 09:25:42 crc kubenswrapper[4934]: I1227 09:25:42.630529 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2lll_c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd/cp-metrics/0.log" Dec 27 09:25:42 crc kubenswrapper[4934]: I1227 09:25:42.634743 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2lll_c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd/controller/0.log" Dec 27 09:25:42 crc kubenswrapper[4934]: I1227 09:25:42.829607 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2lll_c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd/kube-rbac-proxy/0.log" Dec 27 09:25:42 crc kubenswrapper[4934]: I1227 09:25:42.862726 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2lll_c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd/frr-metrics/0.log" Dec 27 09:25:42 crc kubenswrapper[4934]: I1227 09:25:42.932515 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2lll_c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd/frr/1.log" Dec 27 09:25:43 crc kubenswrapper[4934]: I1227 09:25:43.030903 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2lll_c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd/kube-rbac-proxy-frr/0.log" Dec 27 09:25:43 crc kubenswrapper[4934]: I1227 09:25:43.083772 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2lll_c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd/reloader/0.log" Dec 27 09:25:43 crc kubenswrapper[4934]: I1227 09:25:43.245796 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7784b6fcf-5bsq9_0134c36f-c43b-4830-848a-3a8690957ee1/frr-k8s-webhook-server/0.log" Dec 27 09:25:43 crc kubenswrapper[4934]: I1227 09:25:43.474548 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5b75b656f5-sqn2z_fa5c19f1-9e47-4c7e-a4cf-28bb6d38b107/manager/0.log" Dec 27 09:25:43 crc kubenswrapper[4934]: I1227 09:25:43.482433 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-8b4577cdd-qcwh8_df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb/webhook-server/1.log" Dec 27 09:25:43 crc kubenswrapper[4934]: I1227 09:25:43.664369 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-8b4577cdd-qcwh8_df8d44b2-c06a-4d4f-8ff8-6bde8f8b4acb/webhook-server/0.log" Dec 27 09:25:43 crc kubenswrapper[4934]: I1227 09:25:43.675761 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-4jwwj_ea01c7df-c5aa-4c86-800b-b6aebafeb8fd/kube-rbac-proxy/0.log" Dec 27 09:25:44 crc kubenswrapper[4934]: I1227 09:25:44.828241 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-4jwwj_ea01c7df-c5aa-4c86-800b-b6aebafeb8fd/speaker/0.log" Dec 27 09:25:44 crc kubenswrapper[4934]: I1227 09:25:44.858402 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-v2lll_c1651ea1-c429-4dfd-b4de-ffd05d3fc4cd/frr/0.log" Dec 27 09:25:45 crc kubenswrapper[4934]: I1227 09:25:45.330008 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 09:25:45 crc kubenswrapper[4934]: I1227 09:25:45.330108 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 09:25:57 crc kubenswrapper[4934]: I1227 09:25:57.809896 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt_2befd093-f28c-4779-b290-eb99f3ef5cc2/util/0.log" Dec 27 09:25:57 crc kubenswrapper[4934]: I1227 09:25:57.987004 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt_2befd093-f28c-4779-b290-eb99f3ef5cc2/pull/0.log" Dec 27 09:25:58 crc kubenswrapper[4934]: I1227 09:25:58.030895 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt_2befd093-f28c-4779-b290-eb99f3ef5cc2/pull/0.log" Dec 27 09:25:58 crc kubenswrapper[4934]: I1227 09:25:58.031210 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt_2befd093-f28c-4779-b290-eb99f3ef5cc2/util/0.log" Dec 27 09:25:58 crc kubenswrapper[4934]: I1227 09:25:58.214844 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt_2befd093-f28c-4779-b290-eb99f3ef5cc2/extract/0.log" Dec 27 09:25:58 crc kubenswrapper[4934]: I1227 09:25:58.226679 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt_2befd093-f28c-4779-b290-eb99f3ef5cc2/util/0.log" Dec 27 09:25:58 crc kubenswrapper[4934]: I1227 09:25:58.236811 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_19f7b28a9b43ae652fc2e0b84ee4ec326dbd0a997d417d0c402b7249a274xlt_2befd093-f28c-4779-b290-eb99f3ef5cc2/pull/0.log" Dec 27 09:25:58 crc kubenswrapper[4934]: I1227 09:25:58.379343 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8_74e9f2f0-b2fa-4574-9378-263ef6468eda/util/0.log" Dec 27 09:25:58 crc kubenswrapper[4934]: I1227 09:25:58.559744 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8_74e9f2f0-b2fa-4574-9378-263ef6468eda/pull/0.log" Dec 27 09:25:58 crc kubenswrapper[4934]: I1227 09:25:58.567372 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8_74e9f2f0-b2fa-4574-9378-263ef6468eda/util/0.log" Dec 27 09:25:58 crc kubenswrapper[4934]: I1227 09:25:58.590541 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8_74e9f2f0-b2fa-4574-9378-263ef6468eda/pull/0.log" Dec 27 09:25:58 crc kubenswrapper[4934]: I1227 09:25:58.762793 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8_74e9f2f0-b2fa-4574-9378-263ef6468eda/util/0.log" Dec 27 09:25:58 crc kubenswrapper[4934]: I1227 09:25:58.780604 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8_74e9f2f0-b2fa-4574-9378-263ef6468eda/pull/0.log" Dec 27 09:25:58 crc kubenswrapper[4934]: I1227 09:25:58.813777 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_40d905839fa7263f1f473fab6e11a9af2a700db4f753f3af512410360bfwqg8_74e9f2f0-b2fa-4574-9378-263ef6468eda/extract/0.log" Dec 27 09:25:58 crc kubenswrapper[4934]: I1227 09:25:58.951250 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x_75746381-e0b6-4ad6-8238-ea5f55cb7f54/util/0.log" Dec 27 09:25:59 crc kubenswrapper[4934]: I1227 09:25:59.168138 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x_75746381-e0b6-4ad6-8238-ea5f55cb7f54/pull/0.log" Dec 27 09:25:59 crc kubenswrapper[4934]: I1227 09:25:59.170041 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x_75746381-e0b6-4ad6-8238-ea5f55cb7f54/util/0.log" Dec 27 09:25:59 crc kubenswrapper[4934]: I1227 09:25:59.170183 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x_75746381-e0b6-4ad6-8238-ea5f55cb7f54/pull/0.log" Dec 27 09:25:59 crc kubenswrapper[4934]: I1227 09:25:59.318682 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x_75746381-e0b6-4ad6-8238-ea5f55cb7f54/pull/0.log" Dec 27 09:25:59 crc kubenswrapper[4934]: I1227 09:25:59.352512 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x_75746381-e0b6-4ad6-8238-ea5f55cb7f54/util/0.log" Dec 27 09:25:59 crc kubenswrapper[4934]: I1227 09:25:59.363057 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5b7fccbebf0e22d2dd769066fa7aaa90fd620c5db34f2af6c91e4319d4zdh7x_75746381-e0b6-4ad6-8238-ea5f55cb7f54/extract/0.log" Dec 27 09:25:59 crc kubenswrapper[4934]: I1227 09:25:59.516541 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp_7f133bd4-74ca-4b6c-bd9a-210102808e22/util/0.log" Dec 27 09:25:59 crc kubenswrapper[4934]: I1227 09:25:59.694092 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp_7f133bd4-74ca-4b6c-bd9a-210102808e22/pull/0.log" Dec 27 09:25:59 crc kubenswrapper[4934]: I1227 09:25:59.697957 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp_7f133bd4-74ca-4b6c-bd9a-210102808e22/util/0.log" Dec 27 09:25:59 crc kubenswrapper[4934]: I1227 09:25:59.725640 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp_7f133bd4-74ca-4b6c-bd9a-210102808e22/pull/0.log" Dec 27 09:25:59 crc kubenswrapper[4934]: I1227 09:25:59.885367 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp_7f133bd4-74ca-4b6c-bd9a-210102808e22/util/0.log" Dec 27 09:25:59 crc kubenswrapper[4934]: I1227 09:25:59.885761 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp_7f133bd4-74ca-4b6c-bd9a-210102808e22/extract/0.log" Dec 27 09:25:59 crc kubenswrapper[4934]: I1227 09:25:59.887517 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98085b0df3808ebec39f9f9529f737144fe2dbcdaa4f334014817c0fa8b67xp_7f133bd4-74ca-4b6c-bd9a-210102808e22/pull/0.log" Dec 27 09:26:00 crc kubenswrapper[4934]: I1227 09:26:00.069396 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh_d1a4dc79-8563-47a8-afc8-fe018f9bb949/util/0.log" Dec 27 09:26:00 crc kubenswrapper[4934]: I1227 09:26:00.241531 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh_d1a4dc79-8563-47a8-afc8-fe018f9bb949/pull/0.log" Dec 27 09:26:00 crc kubenswrapper[4934]: I1227 09:26:00.241612 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh_d1a4dc79-8563-47a8-afc8-fe018f9bb949/pull/0.log" Dec 27 09:26:00 crc kubenswrapper[4934]: I1227 09:26:00.258591 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh_d1a4dc79-8563-47a8-afc8-fe018f9bb949/util/0.log" Dec 27 09:26:00 crc kubenswrapper[4934]: I1227 09:26:00.418631 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh_d1a4dc79-8563-47a8-afc8-fe018f9bb949/pull/0.log" Dec 27 09:26:00 crc kubenswrapper[4934]: I1227 09:26:00.444355 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh_d1a4dc79-8563-47a8-afc8-fe018f9bb949/util/0.log" Dec 27 09:26:00 crc kubenswrapper[4934]: I1227 09:26:00.445374 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08ps6hh_d1a4dc79-8563-47a8-afc8-fe018f9bb949/extract/0.log" Dec 27 09:26:00 crc kubenswrapper[4934]: I1227 09:26:00.593705 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lbfd2_0661ef0e-af9d-41ba-8059-2e878b11b042/extract-utilities/0.log" Dec 27 09:26:00 crc kubenswrapper[4934]: I1227 09:26:00.799460 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lbfd2_0661ef0e-af9d-41ba-8059-2e878b11b042/extract-content/0.log" Dec 27 09:26:00 crc kubenswrapper[4934]: I1227 09:26:00.799465 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lbfd2_0661ef0e-af9d-41ba-8059-2e878b11b042/extract-utilities/0.log" Dec 27 09:26:00 crc kubenswrapper[4934]: I1227 09:26:00.829754 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lbfd2_0661ef0e-af9d-41ba-8059-2e878b11b042/extract-content/0.log" Dec 27 09:26:00 crc kubenswrapper[4934]: I1227 09:26:00.996910 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lbfd2_0661ef0e-af9d-41ba-8059-2e878b11b042/extract-utilities/0.log" Dec 27 09:26:01 crc kubenswrapper[4934]: I1227 09:26:01.013111 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lbfd2_0661ef0e-af9d-41ba-8059-2e878b11b042/extract-content/0.log" Dec 27 09:26:01 crc kubenswrapper[4934]: I1227 09:26:01.257876 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sz4sq_72569b30-4bc6-4f58-b62e-c1a0ef0961d6/extract-utilities/0.log" Dec 27 09:26:01 crc kubenswrapper[4934]: I1227 09:26:01.497357 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sz4sq_72569b30-4bc6-4f58-b62e-c1a0ef0961d6/extract-utilities/0.log" Dec 27 09:26:01 crc kubenswrapper[4934]: I1227 09:26:01.527817 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sz4sq_72569b30-4bc6-4f58-b62e-c1a0ef0961d6/extract-content/0.log" Dec 27 09:26:01 crc kubenswrapper[4934]: I1227 09:26:01.532749 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sz4sq_72569b30-4bc6-4f58-b62e-c1a0ef0961d6/extract-content/0.log" Dec 27 09:26:01 crc kubenswrapper[4934]: I1227 09:26:01.765354 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sz4sq_72569b30-4bc6-4f58-b62e-c1a0ef0961d6/extract-content/0.log" Dec 27 09:26:01 crc kubenswrapper[4934]: I1227 09:26:01.778462 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sz4sq_72569b30-4bc6-4f58-b62e-c1a0ef0961d6/extract-utilities/0.log" Dec 27 09:26:01 crc kubenswrapper[4934]: I1227 09:26:01.959981 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-5jzhr_a4612dd4-0e0d-4c38-8da5-8bc30dee7c12/marketplace-operator/1.log" Dec 27 09:26:01 crc kubenswrapper[4934]: I1227 09:26:01.964796 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-5jzhr_a4612dd4-0e0d-4c38-8da5-8bc30dee7c12/marketplace-operator/0.log" Dec 27 09:26:02 crc kubenswrapper[4934]: I1227 09:26:02.069071 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lbfd2_0661ef0e-af9d-41ba-8059-2e878b11b042/registry-server/0.log" Dec 27 09:26:02 crc kubenswrapper[4934]: I1227 09:26:02.179816 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7rs6w_ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3/extract-utilities/0.log" Dec 27 09:26:02 crc kubenswrapper[4934]: I1227 09:26:02.388417 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7rs6w_ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3/extract-content/0.log" Dec 27 09:26:02 crc kubenswrapper[4934]: I1227 09:26:02.403417 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7rs6w_ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3/extract-content/0.log" Dec 27 09:26:02 crc kubenswrapper[4934]: I1227 09:26:02.438327 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7rs6w_ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3/extract-utilities/0.log" Dec 27 09:26:02 crc kubenswrapper[4934]: I1227 09:26:02.669038 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7rs6w_ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3/extract-content/0.log" Dec 27 09:26:02 crc kubenswrapper[4934]: I1227 09:26:02.725979 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7rs6w_ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3/extract-utilities/0.log" Dec 27 09:26:02 crc kubenswrapper[4934]: I1227 09:26:02.905596 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qtq22_2ee60051-8eec-4da3-ae78-b409c5cdf9f3/extract-utilities/0.log" Dec 27 09:26:03 crc kubenswrapper[4934]: I1227 09:26:03.034479 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-sz4sq_72569b30-4bc6-4f58-b62e-c1a0ef0961d6/registry-server/0.log" Dec 27 09:26:03 crc kubenswrapper[4934]: I1227 09:26:03.116763 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qtq22_2ee60051-8eec-4da3-ae78-b409c5cdf9f3/extract-utilities/0.log" Dec 27 09:26:03 crc kubenswrapper[4934]: I1227 09:26:03.139812 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-7rs6w_ec69ef3c-f87e-4b0d-ba44-9067f8e4a2b3/registry-server/0.log" Dec 27 09:26:03 crc kubenswrapper[4934]: I1227 09:26:03.207909 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qtq22_2ee60051-8eec-4da3-ae78-b409c5cdf9f3/extract-content/0.log" Dec 27 09:26:03 crc kubenswrapper[4934]: I1227 09:26:03.209727 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qtq22_2ee60051-8eec-4da3-ae78-b409c5cdf9f3/extract-content/0.log" Dec 27 09:26:03 crc kubenswrapper[4934]: I1227 09:26:03.328695 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qtq22_2ee60051-8eec-4da3-ae78-b409c5cdf9f3/extract-content/0.log" Dec 27 09:26:03 crc kubenswrapper[4934]: I1227 09:26:03.333455 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qtq22_2ee60051-8eec-4da3-ae78-b409c5cdf9f3/extract-utilities/0.log" Dec 27 09:26:04 crc kubenswrapper[4934]: I1227 09:26:04.275314 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qtq22_2ee60051-8eec-4da3-ae78-b409c5cdf9f3/registry-server/0.log" Dec 27 09:26:15 crc kubenswrapper[4934]: I1227 09:26:15.330426 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 09:26:15 crc kubenswrapper[4934]: I1227 09:26:15.331018 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 09:26:15 crc kubenswrapper[4934]: I1227 09:26:15.331071 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 09:26:15 crc kubenswrapper[4934]: I1227 09:26:15.332253 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f3e74ce13c29f6896d91ffd6e9150b3e5f47b60d6ed52c857f27a4eb1e7598d0"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 09:26:15 crc kubenswrapper[4934]: I1227 09:26:15.332313 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://f3e74ce13c29f6896d91ffd6e9150b3e5f47b60d6ed52c857f27a4eb1e7598d0" gracePeriod=600 Dec 27 09:26:16 crc kubenswrapper[4934]: I1227 09:26:16.231256 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="f3e74ce13c29f6896d91ffd6e9150b3e5f47b60d6ed52c857f27a4eb1e7598d0" exitCode=0 Dec 27 09:26:16 crc kubenswrapper[4934]: I1227 09:26:16.231338 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"f3e74ce13c29f6896d91ffd6e9150b3e5f47b60d6ed52c857f27a4eb1e7598d0"} Dec 27 09:26:16 crc kubenswrapper[4934]: I1227 09:26:16.231691 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerStarted","Data":"9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5"} Dec 27 09:26:16 crc kubenswrapper[4934]: I1227 09:26:16.231712 4934 scope.go:117] "RemoveContainer" containerID="7851497d6dcd3d6e25f21d9f6b25ef27c300c849cc47b3405cd09735027ec41c" Dec 27 09:26:16 crc kubenswrapper[4934]: I1227 09:26:16.542324 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-8llj6_be52e823-2384-41da-b213-85e7fe61250e/prometheus-operator/0.log" Dec 27 09:26:16 crc kubenswrapper[4934]: I1227 09:26:16.713026 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7499b5c98c-j52vn_b86f501c-2f1d-4131-820b-c9b71c4973dc/prometheus-operator-admission-webhook/0.log" Dec 27 09:26:16 crc kubenswrapper[4934]: I1227 09:26:16.778375 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-7499b5c98c-s9bc8_084dcdaf-73d5-45a7-9d98-a1dbcd76db41/prometheus-operator-admission-webhook/0.log" Dec 27 09:26:16 crc kubenswrapper[4934]: I1227 09:26:16.919232 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-lnknm_4575d285-879f-4cc3-a9b9-406dc96e852d/operator/1.log" Dec 27 09:26:16 crc kubenswrapper[4934]: I1227 09:26:16.978425 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-lnknm_4575d285-879f-4cc3-a9b9-406dc96e852d/operator/0.log" Dec 27 09:26:17 crc kubenswrapper[4934]: I1227 09:26:17.101244 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-ui-dashboards-66cbf594b5-swd4w_355bcd5e-7852-477f-8e4b-b03847566cc6/observability-ui-dashboards/0.log" Dec 27 09:26:17 crc kubenswrapper[4934]: I1227 09:26:17.179848 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-tx6z5_f97e8d86-1e9e-4c5a-a164-2600193c576b/perses-operator/0.log" Dec 27 09:26:30 crc kubenswrapper[4934]: I1227 09:26:30.783102 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-77c7786d5-gz9d9_00799e47-9b94-49bf-8bd1-d6bb5036285a/kube-rbac-proxy/0.log" Dec 27 09:26:30 crc kubenswrapper[4934]: I1227 09:26:30.817318 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-77c7786d5-gz9d9_00799e47-9b94-49bf-8bd1-d6bb5036285a/manager/1.log" Dec 27 09:26:30 crc kubenswrapper[4934]: I1227 09:26:30.839464 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-77c7786d5-gz9d9_00799e47-9b94-49bf-8bd1-d6bb5036285a/manager/0.log" Dec 27 09:26:51 crc kubenswrapper[4934]: E1227 09:26:51.605998 4934 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.13:52142->38.102.83.13:43999: write tcp 38.102.83.13:52142->38.102.83.13:43999: write: broken pipe Dec 27 09:27:53 crc kubenswrapper[4934]: I1227 09:27:53.649771 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tf5k9"] Dec 27 09:27:53 crc kubenswrapper[4934]: E1227 09:27:53.651915 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee2dcc57-b167-4885-8a2a-b05979c842f7" containerName="container-00" Dec 27 09:27:53 crc kubenswrapper[4934]: I1227 09:27:53.651938 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee2dcc57-b167-4885-8a2a-b05979c842f7" containerName="container-00" Dec 27 09:27:53 crc kubenswrapper[4934]: I1227 09:27:53.652317 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee2dcc57-b167-4885-8a2a-b05979c842f7" containerName="container-00" Dec 27 09:27:53 crc kubenswrapper[4934]: I1227 09:27:53.655839 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tf5k9" Dec 27 09:27:53 crc kubenswrapper[4934]: I1227 09:27:53.690553 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tf5k9"] Dec 27 09:27:53 crc kubenswrapper[4934]: I1227 09:27:53.778813 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54514998-5b8e-49af-865b-a214a575b07e-catalog-content\") pod \"community-operators-tf5k9\" (UID: \"54514998-5b8e-49af-865b-a214a575b07e\") " pod="openshift-marketplace/community-operators-tf5k9" Dec 27 09:27:53 crc kubenswrapper[4934]: I1227 09:27:53.778866 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54514998-5b8e-49af-865b-a214a575b07e-utilities\") pod \"community-operators-tf5k9\" (UID: \"54514998-5b8e-49af-865b-a214a575b07e\") " pod="openshift-marketplace/community-operators-tf5k9" Dec 27 09:27:53 crc kubenswrapper[4934]: I1227 09:27:53.779516 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsfzb\" (UniqueName: \"kubernetes.io/projected/54514998-5b8e-49af-865b-a214a575b07e-kube-api-access-zsfzb\") pod \"community-operators-tf5k9\" (UID: \"54514998-5b8e-49af-865b-a214a575b07e\") " pod="openshift-marketplace/community-operators-tf5k9" Dec 27 09:27:53 crc kubenswrapper[4934]: I1227 09:27:53.881983 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54514998-5b8e-49af-865b-a214a575b07e-catalog-content\") pod \"community-operators-tf5k9\" (UID: \"54514998-5b8e-49af-865b-a214a575b07e\") " pod="openshift-marketplace/community-operators-tf5k9" Dec 27 09:27:53 crc kubenswrapper[4934]: I1227 09:27:53.882033 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54514998-5b8e-49af-865b-a214a575b07e-utilities\") pod \"community-operators-tf5k9\" (UID: \"54514998-5b8e-49af-865b-a214a575b07e\") " pod="openshift-marketplace/community-operators-tf5k9" Dec 27 09:27:53 crc kubenswrapper[4934]: I1227 09:27:53.882298 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsfzb\" (UniqueName: \"kubernetes.io/projected/54514998-5b8e-49af-865b-a214a575b07e-kube-api-access-zsfzb\") pod \"community-operators-tf5k9\" (UID: \"54514998-5b8e-49af-865b-a214a575b07e\") " pod="openshift-marketplace/community-operators-tf5k9" Dec 27 09:27:53 crc kubenswrapper[4934]: I1227 09:27:53.882673 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54514998-5b8e-49af-865b-a214a575b07e-catalog-content\") pod \"community-operators-tf5k9\" (UID: \"54514998-5b8e-49af-865b-a214a575b07e\") " pod="openshift-marketplace/community-operators-tf5k9" Dec 27 09:27:53 crc kubenswrapper[4934]: I1227 09:27:53.882777 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54514998-5b8e-49af-865b-a214a575b07e-utilities\") pod \"community-operators-tf5k9\" (UID: \"54514998-5b8e-49af-865b-a214a575b07e\") " pod="openshift-marketplace/community-operators-tf5k9" Dec 27 09:27:53 crc kubenswrapper[4934]: I1227 09:27:53.903126 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsfzb\" (UniqueName: \"kubernetes.io/projected/54514998-5b8e-49af-865b-a214a575b07e-kube-api-access-zsfzb\") pod \"community-operators-tf5k9\" (UID: \"54514998-5b8e-49af-865b-a214a575b07e\") " pod="openshift-marketplace/community-operators-tf5k9" Dec 27 09:27:53 crc kubenswrapper[4934]: I1227 09:27:53.982502 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tf5k9" Dec 27 09:27:54 crc kubenswrapper[4934]: I1227 09:27:54.646627 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tf5k9"] Dec 27 09:27:54 crc kubenswrapper[4934]: W1227 09:27:54.649341 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54514998_5b8e_49af_865b_a214a575b07e.slice/crio-f2abc277f2d04f7282f9bf86401bb04fb9bd5443647d9143db41d861a8c55370 WatchSource:0}: Error finding container f2abc277f2d04f7282f9bf86401bb04fb9bd5443647d9143db41d861a8c55370: Status 404 returned error can't find the container with id f2abc277f2d04f7282f9bf86401bb04fb9bd5443647d9143db41d861a8c55370 Dec 27 09:27:54 crc kubenswrapper[4934]: I1227 09:27:54.690642 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tf5k9" event={"ID":"54514998-5b8e-49af-865b-a214a575b07e","Type":"ContainerStarted","Data":"f2abc277f2d04f7282f9bf86401bb04fb9bd5443647d9143db41d861a8c55370"} Dec 27 09:27:55 crc kubenswrapper[4934]: I1227 09:27:55.709012 4934 generic.go:334] "Generic (PLEG): container finished" podID="54514998-5b8e-49af-865b-a214a575b07e" containerID="24bf05ac09c9192a9358f2ee9ed96b1f11ac9ff5153ad6b49344ce91d32ff836" exitCode=0 Dec 27 09:27:55 crc kubenswrapper[4934]: I1227 09:27:55.710215 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tf5k9" event={"ID":"54514998-5b8e-49af-865b-a214a575b07e","Type":"ContainerDied","Data":"24bf05ac09c9192a9358f2ee9ed96b1f11ac9ff5153ad6b49344ce91d32ff836"} Dec 27 09:27:55 crc kubenswrapper[4934]: I1227 09:27:55.716194 4934 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 27 09:27:56 crc kubenswrapper[4934]: I1227 09:27:56.725141 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tf5k9" event={"ID":"54514998-5b8e-49af-865b-a214a575b07e","Type":"ContainerStarted","Data":"ae27a1403e302faaa82af5417a983394f4a374c67dbef4c68a066bfd3fa81fb1"} Dec 27 09:27:57 crc kubenswrapper[4934]: I1227 09:27:57.743310 4934 generic.go:334] "Generic (PLEG): container finished" podID="54514998-5b8e-49af-865b-a214a575b07e" containerID="ae27a1403e302faaa82af5417a983394f4a374c67dbef4c68a066bfd3fa81fb1" exitCode=0 Dec 27 09:27:57 crc kubenswrapper[4934]: I1227 09:27:57.743425 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tf5k9" event={"ID":"54514998-5b8e-49af-865b-a214a575b07e","Type":"ContainerDied","Data":"ae27a1403e302faaa82af5417a983394f4a374c67dbef4c68a066bfd3fa81fb1"} Dec 27 09:27:58 crc kubenswrapper[4934]: I1227 09:27:58.761922 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tf5k9" event={"ID":"54514998-5b8e-49af-865b-a214a575b07e","Type":"ContainerStarted","Data":"b2ed0ae9c5ac28b4b5779a4c70747c1e99fb27a5b4c66ea1b7f40a859e6ad71b"} Dec 27 09:27:58 crc kubenswrapper[4934]: I1227 09:27:58.785074 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tf5k9" podStartSLOduration=3.256062881 podStartE2EDuration="5.785050035s" podCreationTimestamp="2025-12-27 09:27:53 +0000 UTC" firstStartedPulling="2025-12-27 09:27:55.713569476 +0000 UTC m=+6336.534010080" lastFinishedPulling="2025-12-27 09:27:58.24255664 +0000 UTC m=+6339.062997234" observedRunningTime="2025-12-27 09:27:58.782874161 +0000 UTC m=+6339.603314765" watchObservedRunningTime="2025-12-27 09:27:58.785050035 +0000 UTC m=+6339.605490669" Dec 27 09:28:03 crc kubenswrapper[4934]: I1227 09:28:03.983292 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tf5k9" Dec 27 09:28:03 crc kubenswrapper[4934]: I1227 09:28:03.983635 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tf5k9" Dec 27 09:28:04 crc kubenswrapper[4934]: I1227 09:28:04.051216 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tf5k9" Dec 27 09:28:04 crc kubenswrapper[4934]: I1227 09:28:04.928071 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tf5k9" Dec 27 09:28:04 crc kubenswrapper[4934]: I1227 09:28:04.978096 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tf5k9"] Dec 27 09:28:06 crc kubenswrapper[4934]: I1227 09:28:06.897462 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tf5k9" podUID="54514998-5b8e-49af-865b-a214a575b07e" containerName="registry-server" containerID="cri-o://b2ed0ae9c5ac28b4b5779a4c70747c1e99fb27a5b4c66ea1b7f40a859e6ad71b" gracePeriod=2 Dec 27 09:28:07 crc kubenswrapper[4934]: I1227 09:28:07.917575 4934 generic.go:334] "Generic (PLEG): container finished" podID="54514998-5b8e-49af-865b-a214a575b07e" containerID="b2ed0ae9c5ac28b4b5779a4c70747c1e99fb27a5b4c66ea1b7f40a859e6ad71b" exitCode=0 Dec 27 09:28:07 crc kubenswrapper[4934]: I1227 09:28:07.917645 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tf5k9" event={"ID":"54514998-5b8e-49af-865b-a214a575b07e","Type":"ContainerDied","Data":"b2ed0ae9c5ac28b4b5779a4c70747c1e99fb27a5b4c66ea1b7f40a859e6ad71b"} Dec 27 09:28:08 crc kubenswrapper[4934]: I1227 09:28:08.043205 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tf5k9" Dec 27 09:28:08 crc kubenswrapper[4934]: I1227 09:28:08.190093 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54514998-5b8e-49af-865b-a214a575b07e-catalog-content\") pod \"54514998-5b8e-49af-865b-a214a575b07e\" (UID: \"54514998-5b8e-49af-865b-a214a575b07e\") " Dec 27 09:28:08 crc kubenswrapper[4934]: I1227 09:28:08.190487 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54514998-5b8e-49af-865b-a214a575b07e-utilities\") pod \"54514998-5b8e-49af-865b-a214a575b07e\" (UID: \"54514998-5b8e-49af-865b-a214a575b07e\") " Dec 27 09:28:08 crc kubenswrapper[4934]: I1227 09:28:08.190561 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsfzb\" (UniqueName: \"kubernetes.io/projected/54514998-5b8e-49af-865b-a214a575b07e-kube-api-access-zsfzb\") pod \"54514998-5b8e-49af-865b-a214a575b07e\" (UID: \"54514998-5b8e-49af-865b-a214a575b07e\") " Dec 27 09:28:08 crc kubenswrapper[4934]: I1227 09:28:08.204811 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54514998-5b8e-49af-865b-a214a575b07e-utilities" (OuterVolumeSpecName: "utilities") pod "54514998-5b8e-49af-865b-a214a575b07e" (UID: "54514998-5b8e-49af-865b-a214a575b07e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:28:08 crc kubenswrapper[4934]: I1227 09:28:08.205926 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54514998-5b8e-49af-865b-a214a575b07e-kube-api-access-zsfzb" (OuterVolumeSpecName: "kube-api-access-zsfzb") pod "54514998-5b8e-49af-865b-a214a575b07e" (UID: "54514998-5b8e-49af-865b-a214a575b07e"). InnerVolumeSpecName "kube-api-access-zsfzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:28:08 crc kubenswrapper[4934]: I1227 09:28:08.284983 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/54514998-5b8e-49af-865b-a214a575b07e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "54514998-5b8e-49af-865b-a214a575b07e" (UID: "54514998-5b8e-49af-865b-a214a575b07e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:28:08 crc kubenswrapper[4934]: I1227 09:28:08.298661 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/54514998-5b8e-49af-865b-a214a575b07e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 09:28:08 crc kubenswrapper[4934]: I1227 09:28:08.298706 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/54514998-5b8e-49af-865b-a214a575b07e-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 09:28:08 crc kubenswrapper[4934]: I1227 09:28:08.298718 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsfzb\" (UniqueName: \"kubernetes.io/projected/54514998-5b8e-49af-865b-a214a575b07e-kube-api-access-zsfzb\") on node \"crc\" DevicePath \"\"" Dec 27 09:28:08 crc kubenswrapper[4934]: I1227 09:28:08.939281 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tf5k9" event={"ID":"54514998-5b8e-49af-865b-a214a575b07e","Type":"ContainerDied","Data":"f2abc277f2d04f7282f9bf86401bb04fb9bd5443647d9143db41d861a8c55370"} Dec 27 09:28:08 crc kubenswrapper[4934]: I1227 09:28:08.939386 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tf5k9" Dec 27 09:28:08 crc kubenswrapper[4934]: I1227 09:28:08.940495 4934 scope.go:117] "RemoveContainer" containerID="b2ed0ae9c5ac28b4b5779a4c70747c1e99fb27a5b4c66ea1b7f40a859e6ad71b" Dec 27 09:28:08 crc kubenswrapper[4934]: I1227 09:28:08.984228 4934 scope.go:117] "RemoveContainer" containerID="ae27a1403e302faaa82af5417a983394f4a374c67dbef4c68a066bfd3fa81fb1" Dec 27 09:28:09 crc kubenswrapper[4934]: I1227 09:28:09.001166 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tf5k9"] Dec 27 09:28:09 crc kubenswrapper[4934]: I1227 09:28:09.023593 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tf5k9"] Dec 27 09:28:09 crc kubenswrapper[4934]: I1227 09:28:09.023786 4934 scope.go:117] "RemoveContainer" containerID="24bf05ac09c9192a9358f2ee9ed96b1f11ac9ff5153ad6b49344ce91d32ff836" Dec 27 09:28:09 crc kubenswrapper[4934]: I1227 09:28:09.483763 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54514998-5b8e-49af-865b-a214a575b07e" path="/var/lib/kubelet/pods/54514998-5b8e-49af-865b-a214a575b07e/volumes" Dec 27 09:28:14 crc kubenswrapper[4934]: I1227 09:28:14.181881 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6mtww"] Dec 27 09:28:14 crc kubenswrapper[4934]: E1227 09:28:14.183341 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54514998-5b8e-49af-865b-a214a575b07e" containerName="extract-content" Dec 27 09:28:14 crc kubenswrapper[4934]: I1227 09:28:14.183367 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="54514998-5b8e-49af-865b-a214a575b07e" containerName="extract-content" Dec 27 09:28:14 crc kubenswrapper[4934]: E1227 09:28:14.183409 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54514998-5b8e-49af-865b-a214a575b07e" containerName="registry-server" Dec 27 09:28:14 crc kubenswrapper[4934]: I1227 09:28:14.183420 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="54514998-5b8e-49af-865b-a214a575b07e" containerName="registry-server" Dec 27 09:28:14 crc kubenswrapper[4934]: E1227 09:28:14.183503 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54514998-5b8e-49af-865b-a214a575b07e" containerName="extract-utilities" Dec 27 09:28:14 crc kubenswrapper[4934]: I1227 09:28:14.183516 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="54514998-5b8e-49af-865b-a214a575b07e" containerName="extract-utilities" Dec 27 09:28:14 crc kubenswrapper[4934]: I1227 09:28:14.183916 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="54514998-5b8e-49af-865b-a214a575b07e" containerName="registry-server" Dec 27 09:28:14 crc kubenswrapper[4934]: I1227 09:28:14.187985 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6mtww" Dec 27 09:28:14 crc kubenswrapper[4934]: I1227 09:28:14.215846 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6mtww"] Dec 27 09:28:14 crc kubenswrapper[4934]: I1227 09:28:14.298840 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k785z\" (UniqueName: \"kubernetes.io/projected/91ceeb1a-a053-4bcf-8072-2796ee877381-kube-api-access-k785z\") pod \"certified-operators-6mtww\" (UID: \"91ceeb1a-a053-4bcf-8072-2796ee877381\") " pod="openshift-marketplace/certified-operators-6mtww" Dec 27 09:28:14 crc kubenswrapper[4934]: I1227 09:28:14.298981 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91ceeb1a-a053-4bcf-8072-2796ee877381-utilities\") pod \"certified-operators-6mtww\" (UID: \"91ceeb1a-a053-4bcf-8072-2796ee877381\") " pod="openshift-marketplace/certified-operators-6mtww" Dec 27 09:28:14 crc kubenswrapper[4934]: I1227 09:28:14.299190 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91ceeb1a-a053-4bcf-8072-2796ee877381-catalog-content\") pod \"certified-operators-6mtww\" (UID: \"91ceeb1a-a053-4bcf-8072-2796ee877381\") " pod="openshift-marketplace/certified-operators-6mtww" Dec 27 09:28:14 crc kubenswrapper[4934]: I1227 09:28:14.402134 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k785z\" (UniqueName: \"kubernetes.io/projected/91ceeb1a-a053-4bcf-8072-2796ee877381-kube-api-access-k785z\") pod \"certified-operators-6mtww\" (UID: \"91ceeb1a-a053-4bcf-8072-2796ee877381\") " pod="openshift-marketplace/certified-operators-6mtww" Dec 27 09:28:14 crc kubenswrapper[4934]: I1227 09:28:14.402335 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91ceeb1a-a053-4bcf-8072-2796ee877381-utilities\") pod \"certified-operators-6mtww\" (UID: \"91ceeb1a-a053-4bcf-8072-2796ee877381\") " pod="openshift-marketplace/certified-operators-6mtww" Dec 27 09:28:14 crc kubenswrapper[4934]: I1227 09:28:14.402572 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91ceeb1a-a053-4bcf-8072-2796ee877381-catalog-content\") pod \"certified-operators-6mtww\" (UID: \"91ceeb1a-a053-4bcf-8072-2796ee877381\") " pod="openshift-marketplace/certified-operators-6mtww" Dec 27 09:28:14 crc kubenswrapper[4934]: I1227 09:28:14.403029 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91ceeb1a-a053-4bcf-8072-2796ee877381-utilities\") pod \"certified-operators-6mtww\" (UID: \"91ceeb1a-a053-4bcf-8072-2796ee877381\") " pod="openshift-marketplace/certified-operators-6mtww" Dec 27 09:28:14 crc kubenswrapper[4934]: I1227 09:28:14.403229 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91ceeb1a-a053-4bcf-8072-2796ee877381-catalog-content\") pod \"certified-operators-6mtww\" (UID: \"91ceeb1a-a053-4bcf-8072-2796ee877381\") " pod="openshift-marketplace/certified-operators-6mtww" Dec 27 09:28:14 crc kubenswrapper[4934]: I1227 09:28:14.425196 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k785z\" (UniqueName: \"kubernetes.io/projected/91ceeb1a-a053-4bcf-8072-2796ee877381-kube-api-access-k785z\") pod \"certified-operators-6mtww\" (UID: \"91ceeb1a-a053-4bcf-8072-2796ee877381\") " pod="openshift-marketplace/certified-operators-6mtww" Dec 27 09:28:14 crc kubenswrapper[4934]: I1227 09:28:14.535546 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6mtww" Dec 27 09:28:15 crc kubenswrapper[4934]: I1227 09:28:15.099833 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6mtww"] Dec 27 09:28:15 crc kubenswrapper[4934]: I1227 09:28:15.332803 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 09:28:15 crc kubenswrapper[4934]: I1227 09:28:15.333122 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 09:28:16 crc kubenswrapper[4934]: I1227 09:28:16.020958 4934 generic.go:334] "Generic (PLEG): container finished" podID="91ceeb1a-a053-4bcf-8072-2796ee877381" containerID="785563125c58bbba3ecf11245d14603cfb775d70ee9db1d2cb29ad4245723c9e" exitCode=0 Dec 27 09:28:16 crc kubenswrapper[4934]: I1227 09:28:16.021025 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mtww" event={"ID":"91ceeb1a-a053-4bcf-8072-2796ee877381","Type":"ContainerDied","Data":"785563125c58bbba3ecf11245d14603cfb775d70ee9db1d2cb29ad4245723c9e"} Dec 27 09:28:16 crc kubenswrapper[4934]: I1227 09:28:16.021054 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mtww" event={"ID":"91ceeb1a-a053-4bcf-8072-2796ee877381","Type":"ContainerStarted","Data":"2bb14a4464610c5f7538ebd939ba617c946a7709b4e2056adabc8f574d415204"} Dec 27 09:28:18 crc kubenswrapper[4934]: I1227 09:28:18.051906 4934 generic.go:334] "Generic (PLEG): container finished" podID="91ceeb1a-a053-4bcf-8072-2796ee877381" containerID="c39bc198886641e6f0b4b9f3b1ce7a71bd353007c72c07eb35faaeec1c3a4e3c" exitCode=0 Dec 27 09:28:18 crc kubenswrapper[4934]: I1227 09:28:18.051986 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mtww" event={"ID":"91ceeb1a-a053-4bcf-8072-2796ee877381","Type":"ContainerDied","Data":"c39bc198886641e6f0b4b9f3b1ce7a71bd353007c72c07eb35faaeec1c3a4e3c"} Dec 27 09:28:19 crc kubenswrapper[4934]: I1227 09:28:19.067459 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mtww" event={"ID":"91ceeb1a-a053-4bcf-8072-2796ee877381","Type":"ContainerStarted","Data":"9b549bfb69a764e8832a99c46f6d6ded0bafcf88591dfd3b179792bdac2f0cf1"} Dec 27 09:28:19 crc kubenswrapper[4934]: I1227 09:28:19.118955 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6mtww" podStartSLOduration=2.435407856 podStartE2EDuration="5.118928763s" podCreationTimestamp="2025-12-27 09:28:14 +0000 UTC" firstStartedPulling="2025-12-27 09:28:16.023928834 +0000 UTC m=+6356.844369468" lastFinishedPulling="2025-12-27 09:28:18.707449741 +0000 UTC m=+6359.527890375" observedRunningTime="2025-12-27 09:28:19.09978491 +0000 UTC m=+6359.920225504" watchObservedRunningTime="2025-12-27 09:28:19.118928763 +0000 UTC m=+6359.939369377" Dec 27 09:28:23 crc kubenswrapper[4934]: I1227 09:28:23.614511 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gdrbf"] Dec 27 09:28:23 crc kubenswrapper[4934]: I1227 09:28:23.617767 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gdrbf" Dec 27 09:28:23 crc kubenswrapper[4934]: I1227 09:28:23.633676 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gdrbf"] Dec 27 09:28:23 crc kubenswrapper[4934]: I1227 09:28:23.664034 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xprhm\" (UniqueName: \"kubernetes.io/projected/c0cd0d57-203f-4d71-b26e-0329526839ed-kube-api-access-xprhm\") pod \"redhat-operators-gdrbf\" (UID: \"c0cd0d57-203f-4d71-b26e-0329526839ed\") " pod="openshift-marketplace/redhat-operators-gdrbf" Dec 27 09:28:23 crc kubenswrapper[4934]: I1227 09:28:23.664200 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0cd0d57-203f-4d71-b26e-0329526839ed-utilities\") pod \"redhat-operators-gdrbf\" (UID: \"c0cd0d57-203f-4d71-b26e-0329526839ed\") " pod="openshift-marketplace/redhat-operators-gdrbf" Dec 27 09:28:23 crc kubenswrapper[4934]: I1227 09:28:23.664500 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0cd0d57-203f-4d71-b26e-0329526839ed-catalog-content\") pod \"redhat-operators-gdrbf\" (UID: \"c0cd0d57-203f-4d71-b26e-0329526839ed\") " pod="openshift-marketplace/redhat-operators-gdrbf" Dec 27 09:28:23 crc kubenswrapper[4934]: I1227 09:28:23.766996 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0cd0d57-203f-4d71-b26e-0329526839ed-catalog-content\") pod \"redhat-operators-gdrbf\" (UID: \"c0cd0d57-203f-4d71-b26e-0329526839ed\") " pod="openshift-marketplace/redhat-operators-gdrbf" Dec 27 09:28:23 crc kubenswrapper[4934]: I1227 09:28:23.767158 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xprhm\" (UniqueName: \"kubernetes.io/projected/c0cd0d57-203f-4d71-b26e-0329526839ed-kube-api-access-xprhm\") pod \"redhat-operators-gdrbf\" (UID: \"c0cd0d57-203f-4d71-b26e-0329526839ed\") " pod="openshift-marketplace/redhat-operators-gdrbf" Dec 27 09:28:23 crc kubenswrapper[4934]: I1227 09:28:23.767224 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0cd0d57-203f-4d71-b26e-0329526839ed-utilities\") pod \"redhat-operators-gdrbf\" (UID: \"c0cd0d57-203f-4d71-b26e-0329526839ed\") " pod="openshift-marketplace/redhat-operators-gdrbf" Dec 27 09:28:23 crc kubenswrapper[4934]: I1227 09:28:23.767532 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0cd0d57-203f-4d71-b26e-0329526839ed-catalog-content\") pod \"redhat-operators-gdrbf\" (UID: \"c0cd0d57-203f-4d71-b26e-0329526839ed\") " pod="openshift-marketplace/redhat-operators-gdrbf" Dec 27 09:28:23 crc kubenswrapper[4934]: I1227 09:28:23.767717 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0cd0d57-203f-4d71-b26e-0329526839ed-utilities\") pod \"redhat-operators-gdrbf\" (UID: \"c0cd0d57-203f-4d71-b26e-0329526839ed\") " pod="openshift-marketplace/redhat-operators-gdrbf" Dec 27 09:28:23 crc kubenswrapper[4934]: I1227 09:28:23.788212 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xprhm\" (UniqueName: \"kubernetes.io/projected/c0cd0d57-203f-4d71-b26e-0329526839ed-kube-api-access-xprhm\") pod \"redhat-operators-gdrbf\" (UID: \"c0cd0d57-203f-4d71-b26e-0329526839ed\") " pod="openshift-marketplace/redhat-operators-gdrbf" Dec 27 09:28:23 crc kubenswrapper[4934]: I1227 09:28:23.936725 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gdrbf" Dec 27 09:28:24 crc kubenswrapper[4934]: I1227 09:28:24.417289 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gdrbf"] Dec 27 09:28:24 crc kubenswrapper[4934]: I1227 09:28:24.536543 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6mtww" Dec 27 09:28:24 crc kubenswrapper[4934]: I1227 09:28:24.536591 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6mtww" Dec 27 09:28:24 crc kubenswrapper[4934]: I1227 09:28:24.604246 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6mtww" Dec 27 09:28:25 crc kubenswrapper[4934]: I1227 09:28:25.158052 4934 generic.go:334] "Generic (PLEG): container finished" podID="c0cd0d57-203f-4d71-b26e-0329526839ed" containerID="2060c39deb3b89715c8c86c3808846efcbaed4abc26dfc1f9aab8ea80d0a65fa" exitCode=0 Dec 27 09:28:25 crc kubenswrapper[4934]: I1227 09:28:25.158131 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdrbf" event={"ID":"c0cd0d57-203f-4d71-b26e-0329526839ed","Type":"ContainerDied","Data":"2060c39deb3b89715c8c86c3808846efcbaed4abc26dfc1f9aab8ea80d0a65fa"} Dec 27 09:28:25 crc kubenswrapper[4934]: I1227 09:28:25.158201 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdrbf" event={"ID":"c0cd0d57-203f-4d71-b26e-0329526839ed","Type":"ContainerStarted","Data":"de588c56b5238ca29b34271584e4e34f893d117f1835ce51028d9724f9759228"} Dec 27 09:28:25 crc kubenswrapper[4934]: I1227 09:28:25.227950 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6mtww" Dec 27 09:28:26 crc kubenswrapper[4934]: I1227 09:28:26.169598 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdrbf" event={"ID":"c0cd0d57-203f-4d71-b26e-0329526839ed","Type":"ContainerStarted","Data":"4379488e8e9da6b76eab50eaf15124dee97387b1acf342547f19840e667b01f1"} Dec 27 09:28:27 crc kubenswrapper[4934]: I1227 09:28:27.586501 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6mtww"] Dec 27 09:28:27 crc kubenswrapper[4934]: I1227 09:28:27.586970 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6mtww" podUID="91ceeb1a-a053-4bcf-8072-2796ee877381" containerName="registry-server" containerID="cri-o://9b549bfb69a764e8832a99c46f6d6ded0bafcf88591dfd3b179792bdac2f0cf1" gracePeriod=2 Dec 27 09:28:28 crc kubenswrapper[4934]: I1227 09:28:28.193038 4934 generic.go:334] "Generic (PLEG): container finished" podID="91ceeb1a-a053-4bcf-8072-2796ee877381" containerID="9b549bfb69a764e8832a99c46f6d6ded0bafcf88591dfd3b179792bdac2f0cf1" exitCode=0 Dec 27 09:28:28 crc kubenswrapper[4934]: I1227 09:28:28.193221 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mtww" event={"ID":"91ceeb1a-a053-4bcf-8072-2796ee877381","Type":"ContainerDied","Data":"9b549bfb69a764e8832a99c46f6d6ded0bafcf88591dfd3b179792bdac2f0cf1"} Dec 27 09:28:28 crc kubenswrapper[4934]: I1227 09:28:28.859076 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6mtww" Dec 27 09:28:28 crc kubenswrapper[4934]: I1227 09:28:28.993845 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91ceeb1a-a053-4bcf-8072-2796ee877381-utilities\") pod \"91ceeb1a-a053-4bcf-8072-2796ee877381\" (UID: \"91ceeb1a-a053-4bcf-8072-2796ee877381\") " Dec 27 09:28:28 crc kubenswrapper[4934]: I1227 09:28:28.994284 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91ceeb1a-a053-4bcf-8072-2796ee877381-catalog-content\") pod \"91ceeb1a-a053-4bcf-8072-2796ee877381\" (UID: \"91ceeb1a-a053-4bcf-8072-2796ee877381\") " Dec 27 09:28:28 crc kubenswrapper[4934]: I1227 09:28:28.994341 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k785z\" (UniqueName: \"kubernetes.io/projected/91ceeb1a-a053-4bcf-8072-2796ee877381-kube-api-access-k785z\") pod \"91ceeb1a-a053-4bcf-8072-2796ee877381\" (UID: \"91ceeb1a-a053-4bcf-8072-2796ee877381\") " Dec 27 09:28:28 crc kubenswrapper[4934]: I1227 09:28:28.994552 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91ceeb1a-a053-4bcf-8072-2796ee877381-utilities" (OuterVolumeSpecName: "utilities") pod "91ceeb1a-a053-4bcf-8072-2796ee877381" (UID: "91ceeb1a-a053-4bcf-8072-2796ee877381"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:28:28 crc kubenswrapper[4934]: I1227 09:28:28.994992 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91ceeb1a-a053-4bcf-8072-2796ee877381-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 09:28:29 crc kubenswrapper[4934]: I1227 09:28:29.001529 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91ceeb1a-a053-4bcf-8072-2796ee877381-kube-api-access-k785z" (OuterVolumeSpecName: "kube-api-access-k785z") pod "91ceeb1a-a053-4bcf-8072-2796ee877381" (UID: "91ceeb1a-a053-4bcf-8072-2796ee877381"). InnerVolumeSpecName "kube-api-access-k785z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:28:29 crc kubenswrapper[4934]: I1227 09:28:29.041969 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91ceeb1a-a053-4bcf-8072-2796ee877381-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91ceeb1a-a053-4bcf-8072-2796ee877381" (UID: "91ceeb1a-a053-4bcf-8072-2796ee877381"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:28:29 crc kubenswrapper[4934]: I1227 09:28:29.097545 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91ceeb1a-a053-4bcf-8072-2796ee877381-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 09:28:29 crc kubenswrapper[4934]: I1227 09:28:29.097830 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k785z\" (UniqueName: \"kubernetes.io/projected/91ceeb1a-a053-4bcf-8072-2796ee877381-kube-api-access-k785z\") on node \"crc\" DevicePath \"\"" Dec 27 09:28:29 crc kubenswrapper[4934]: I1227 09:28:29.206017 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6mtww" Dec 27 09:28:29 crc kubenswrapper[4934]: I1227 09:28:29.206029 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6mtww" event={"ID":"91ceeb1a-a053-4bcf-8072-2796ee877381","Type":"ContainerDied","Data":"2bb14a4464610c5f7538ebd939ba617c946a7709b4e2056adabc8f574d415204"} Dec 27 09:28:29 crc kubenswrapper[4934]: I1227 09:28:29.206104 4934 scope.go:117] "RemoveContainer" containerID="9b549bfb69a764e8832a99c46f6d6ded0bafcf88591dfd3b179792bdac2f0cf1" Dec 27 09:28:29 crc kubenswrapper[4934]: I1227 09:28:29.209132 4934 generic.go:334] "Generic (PLEG): container finished" podID="c0cd0d57-203f-4d71-b26e-0329526839ed" containerID="4379488e8e9da6b76eab50eaf15124dee97387b1acf342547f19840e667b01f1" exitCode=0 Dec 27 09:28:29 crc kubenswrapper[4934]: I1227 09:28:29.209174 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdrbf" event={"ID":"c0cd0d57-203f-4d71-b26e-0329526839ed","Type":"ContainerDied","Data":"4379488e8e9da6b76eab50eaf15124dee97387b1acf342547f19840e667b01f1"} Dec 27 09:28:29 crc kubenswrapper[4934]: I1227 09:28:29.229834 4934 scope.go:117] "RemoveContainer" containerID="c39bc198886641e6f0b4b9f3b1ce7a71bd353007c72c07eb35faaeec1c3a4e3c" Dec 27 09:28:29 crc kubenswrapper[4934]: I1227 09:28:29.263019 4934 scope.go:117] "RemoveContainer" containerID="785563125c58bbba3ecf11245d14603cfb775d70ee9db1d2cb29ad4245723c9e" Dec 27 09:28:29 crc kubenswrapper[4934]: I1227 09:28:29.268724 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6mtww"] Dec 27 09:28:29 crc kubenswrapper[4934]: I1227 09:28:29.280535 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6mtww"] Dec 27 09:28:29 crc kubenswrapper[4934]: I1227 09:28:29.481418 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91ceeb1a-a053-4bcf-8072-2796ee877381" path="/var/lib/kubelet/pods/91ceeb1a-a053-4bcf-8072-2796ee877381/volumes" Dec 27 09:28:30 crc kubenswrapper[4934]: I1227 09:28:30.221015 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdrbf" event={"ID":"c0cd0d57-203f-4d71-b26e-0329526839ed","Type":"ContainerStarted","Data":"dcc5d8adc2a50ddeae4081e4d0e95c36c3d8895302463f75f298a80479bf813a"} Dec 27 09:28:30 crc kubenswrapper[4934]: I1227 09:28:30.255622 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gdrbf" podStartSLOduration=2.770588588 podStartE2EDuration="7.255600366s" podCreationTimestamp="2025-12-27 09:28:23 +0000 UTC" firstStartedPulling="2025-12-27 09:28:25.161008193 +0000 UTC m=+6365.981448787" lastFinishedPulling="2025-12-27 09:28:29.646019971 +0000 UTC m=+6370.466460565" observedRunningTime="2025-12-27 09:28:30.243877036 +0000 UTC m=+6371.064317630" watchObservedRunningTime="2025-12-27 09:28:30.255600366 +0000 UTC m=+6371.076040960" Dec 27 09:28:33 crc kubenswrapper[4934]: I1227 09:28:33.944133 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gdrbf" Dec 27 09:28:33 crc kubenswrapper[4934]: I1227 09:28:33.945657 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gdrbf" Dec 27 09:28:35 crc kubenswrapper[4934]: I1227 09:28:35.005897 4934 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gdrbf" podUID="c0cd0d57-203f-4d71-b26e-0329526839ed" containerName="registry-server" probeResult="failure" output=< Dec 27 09:28:35 crc kubenswrapper[4934]: timeout: failed to connect service ":50051" within 1s Dec 27 09:28:35 crc kubenswrapper[4934]: > Dec 27 09:28:36 crc kubenswrapper[4934]: I1227 09:28:36.290402 4934 generic.go:334] "Generic (PLEG): container finished" podID="e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f" containerID="44b230a08f0db8ff213b6452d3159f62e7835d4f9033dfd4649d72ae1396c7d9" exitCode=0 Dec 27 09:28:36 crc kubenswrapper[4934]: I1227 09:28:36.290482 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-c8rzb/must-gather-f4p5q" event={"ID":"e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f","Type":"ContainerDied","Data":"44b230a08f0db8ff213b6452d3159f62e7835d4f9033dfd4649d72ae1396c7d9"} Dec 27 09:28:36 crc kubenswrapper[4934]: I1227 09:28:36.291502 4934 scope.go:117] "RemoveContainer" containerID="44b230a08f0db8ff213b6452d3159f62e7835d4f9033dfd4649d72ae1396c7d9" Dec 27 09:28:36 crc kubenswrapper[4934]: I1227 09:28:36.470858 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-c8rzb_must-gather-f4p5q_e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f/gather/0.log" Dec 27 09:28:38 crc kubenswrapper[4934]: I1227 09:28:38.952555 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rztqh"] Dec 27 09:28:38 crc kubenswrapper[4934]: E1227 09:28:38.953637 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91ceeb1a-a053-4bcf-8072-2796ee877381" containerName="registry-server" Dec 27 09:28:38 crc kubenswrapper[4934]: I1227 09:28:38.953651 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="91ceeb1a-a053-4bcf-8072-2796ee877381" containerName="registry-server" Dec 27 09:28:38 crc kubenswrapper[4934]: E1227 09:28:38.953705 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91ceeb1a-a053-4bcf-8072-2796ee877381" containerName="extract-content" Dec 27 09:28:38 crc kubenswrapper[4934]: I1227 09:28:38.953711 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="91ceeb1a-a053-4bcf-8072-2796ee877381" containerName="extract-content" Dec 27 09:28:38 crc kubenswrapper[4934]: E1227 09:28:38.953741 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91ceeb1a-a053-4bcf-8072-2796ee877381" containerName="extract-utilities" Dec 27 09:28:38 crc kubenswrapper[4934]: I1227 09:28:38.953747 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="91ceeb1a-a053-4bcf-8072-2796ee877381" containerName="extract-utilities" Dec 27 09:28:38 crc kubenswrapper[4934]: I1227 09:28:38.953980 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="91ceeb1a-a053-4bcf-8072-2796ee877381" containerName="registry-server" Dec 27 09:28:38 crc kubenswrapper[4934]: I1227 09:28:38.955898 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rztqh" Dec 27 09:28:38 crc kubenswrapper[4934]: I1227 09:28:38.965789 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rztqh"] Dec 27 09:28:39 crc kubenswrapper[4934]: I1227 09:28:39.043942 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a602e20a-d41c-490c-af13-3fd0bffd4519-utilities\") pod \"redhat-marketplace-rztqh\" (UID: \"a602e20a-d41c-490c-af13-3fd0bffd4519\") " pod="openshift-marketplace/redhat-marketplace-rztqh" Dec 27 09:28:39 crc kubenswrapper[4934]: I1227 09:28:39.044017 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qdnr\" (UniqueName: \"kubernetes.io/projected/a602e20a-d41c-490c-af13-3fd0bffd4519-kube-api-access-8qdnr\") pod \"redhat-marketplace-rztqh\" (UID: \"a602e20a-d41c-490c-af13-3fd0bffd4519\") " pod="openshift-marketplace/redhat-marketplace-rztqh" Dec 27 09:28:39 crc kubenswrapper[4934]: I1227 09:28:39.044540 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a602e20a-d41c-490c-af13-3fd0bffd4519-catalog-content\") pod \"redhat-marketplace-rztqh\" (UID: \"a602e20a-d41c-490c-af13-3fd0bffd4519\") " pod="openshift-marketplace/redhat-marketplace-rztqh" Dec 27 09:28:39 crc kubenswrapper[4934]: I1227 09:28:39.147383 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qdnr\" (UniqueName: \"kubernetes.io/projected/a602e20a-d41c-490c-af13-3fd0bffd4519-kube-api-access-8qdnr\") pod \"redhat-marketplace-rztqh\" (UID: \"a602e20a-d41c-490c-af13-3fd0bffd4519\") " pod="openshift-marketplace/redhat-marketplace-rztqh" Dec 27 09:28:39 crc kubenswrapper[4934]: I1227 09:28:39.147618 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a602e20a-d41c-490c-af13-3fd0bffd4519-catalog-content\") pod \"redhat-marketplace-rztqh\" (UID: \"a602e20a-d41c-490c-af13-3fd0bffd4519\") " pod="openshift-marketplace/redhat-marketplace-rztqh" Dec 27 09:28:39 crc kubenswrapper[4934]: I1227 09:28:39.147752 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a602e20a-d41c-490c-af13-3fd0bffd4519-utilities\") pod \"redhat-marketplace-rztqh\" (UID: \"a602e20a-d41c-490c-af13-3fd0bffd4519\") " pod="openshift-marketplace/redhat-marketplace-rztqh" Dec 27 09:28:39 crc kubenswrapper[4934]: I1227 09:28:39.148352 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a602e20a-d41c-490c-af13-3fd0bffd4519-catalog-content\") pod \"redhat-marketplace-rztqh\" (UID: \"a602e20a-d41c-490c-af13-3fd0bffd4519\") " pod="openshift-marketplace/redhat-marketplace-rztqh" Dec 27 09:28:39 crc kubenswrapper[4934]: I1227 09:28:39.148376 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a602e20a-d41c-490c-af13-3fd0bffd4519-utilities\") pod \"redhat-marketplace-rztqh\" (UID: \"a602e20a-d41c-490c-af13-3fd0bffd4519\") " pod="openshift-marketplace/redhat-marketplace-rztqh" Dec 27 09:28:39 crc kubenswrapper[4934]: I1227 09:28:39.169653 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qdnr\" (UniqueName: \"kubernetes.io/projected/a602e20a-d41c-490c-af13-3fd0bffd4519-kube-api-access-8qdnr\") pod \"redhat-marketplace-rztqh\" (UID: \"a602e20a-d41c-490c-af13-3fd0bffd4519\") " pod="openshift-marketplace/redhat-marketplace-rztqh" Dec 27 09:28:39 crc kubenswrapper[4934]: I1227 09:28:39.280770 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rztqh" Dec 27 09:28:39 crc kubenswrapper[4934]: W1227 09:28:39.807964 4934 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda602e20a_d41c_490c_af13_3fd0bffd4519.slice/crio-cdbb8f789ade948bb18a641ac14bdc13e6be3640393506bbecfba26b586f1297 WatchSource:0}: Error finding container cdbb8f789ade948bb18a641ac14bdc13e6be3640393506bbecfba26b586f1297: Status 404 returned error can't find the container with id cdbb8f789ade948bb18a641ac14bdc13e6be3640393506bbecfba26b586f1297 Dec 27 09:28:39 crc kubenswrapper[4934]: I1227 09:28:39.809562 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rztqh"] Dec 27 09:28:40 crc kubenswrapper[4934]: I1227 09:28:40.352893 4934 generic.go:334] "Generic (PLEG): container finished" podID="a602e20a-d41c-490c-af13-3fd0bffd4519" containerID="ed84b2b9f94ee71e5f55702f2589e1a3354c493288bd539faf2c0f909e37ce6e" exitCode=0 Dec 27 09:28:40 crc kubenswrapper[4934]: I1227 09:28:40.353978 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rztqh" event={"ID":"a602e20a-d41c-490c-af13-3fd0bffd4519","Type":"ContainerDied","Data":"ed84b2b9f94ee71e5f55702f2589e1a3354c493288bd539faf2c0f909e37ce6e"} Dec 27 09:28:40 crc kubenswrapper[4934]: I1227 09:28:40.354103 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rztqh" event={"ID":"a602e20a-d41c-490c-af13-3fd0bffd4519","Type":"ContainerStarted","Data":"cdbb8f789ade948bb18a641ac14bdc13e6be3640393506bbecfba26b586f1297"} Dec 27 09:28:41 crc kubenswrapper[4934]: I1227 09:28:41.366151 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rztqh" event={"ID":"a602e20a-d41c-490c-af13-3fd0bffd4519","Type":"ContainerStarted","Data":"8673421efe4027673d3c7a1ce8b3bfc4e08828d32bf5188b02f8f232c6dbe67b"} Dec 27 09:28:42 crc kubenswrapper[4934]: I1227 09:28:42.380679 4934 generic.go:334] "Generic (PLEG): container finished" podID="a602e20a-d41c-490c-af13-3fd0bffd4519" containerID="8673421efe4027673d3c7a1ce8b3bfc4e08828d32bf5188b02f8f232c6dbe67b" exitCode=0 Dec 27 09:28:42 crc kubenswrapper[4934]: I1227 09:28:42.380749 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rztqh" event={"ID":"a602e20a-d41c-490c-af13-3fd0bffd4519","Type":"ContainerDied","Data":"8673421efe4027673d3c7a1ce8b3bfc4e08828d32bf5188b02f8f232c6dbe67b"} Dec 27 09:28:43 crc kubenswrapper[4934]: I1227 09:28:43.391965 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rztqh" event={"ID":"a602e20a-d41c-490c-af13-3fd0bffd4519","Type":"ContainerStarted","Data":"4d9e1a8088a7e48cc784ca16a921024ec63628f3d00492d287578808bd6713d1"} Dec 27 09:28:43 crc kubenswrapper[4934]: I1227 09:28:43.428698 4934 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rztqh" podStartSLOduration=2.704027686 podStartE2EDuration="5.428671021s" podCreationTimestamp="2025-12-27 09:28:38 +0000 UTC" firstStartedPulling="2025-12-27 09:28:40.355485131 +0000 UTC m=+6381.175925725" lastFinishedPulling="2025-12-27 09:28:43.080128456 +0000 UTC m=+6383.900569060" observedRunningTime="2025-12-27 09:28:43.412716586 +0000 UTC m=+6384.233157180" watchObservedRunningTime="2025-12-27 09:28:43.428671021 +0000 UTC m=+6384.249111645" Dec 27 09:28:43 crc kubenswrapper[4934]: I1227 09:28:43.987635 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gdrbf" Dec 27 09:28:44 crc kubenswrapper[4934]: I1227 09:28:44.040318 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gdrbf" Dec 27 09:28:45 crc kubenswrapper[4934]: I1227 09:28:45.330279 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 09:28:45 crc kubenswrapper[4934]: I1227 09:28:45.330834 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 09:28:45 crc kubenswrapper[4934]: I1227 09:28:45.659845 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-c8rzb/must-gather-f4p5q"] Dec 27 09:28:45 crc kubenswrapper[4934]: I1227 09:28:45.660137 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-c8rzb/must-gather-f4p5q" podUID="e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f" containerName="copy" containerID="cri-o://0ce63c7ad4fb139ab4564105fa1322b7292f8c7a0f12262d42f4fcab821aa1ec" gracePeriod=2 Dec 27 09:28:45 crc kubenswrapper[4934]: I1227 09:28:45.676256 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-c8rzb/must-gather-f4p5q"] Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.138169 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gdrbf"] Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.138677 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gdrbf" podUID="c0cd0d57-203f-4d71-b26e-0329526839ed" containerName="registry-server" containerID="cri-o://dcc5d8adc2a50ddeae4081e4d0e95c36c3d8895302463f75f298a80479bf813a" gracePeriod=2 Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.356438 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-c8rzb_must-gather-f4p5q_e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f/copy/0.log" Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.356892 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8rzb/must-gather-f4p5q" Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.430243 4934 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-c8rzb_must-gather-f4p5q_e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f/copy/0.log" Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.431212 4934 generic.go:334] "Generic (PLEG): container finished" podID="e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f" containerID="0ce63c7ad4fb139ab4564105fa1322b7292f8c7a0f12262d42f4fcab821aa1ec" exitCode=143 Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.431262 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-c8rzb/must-gather-f4p5q" Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.431294 4934 scope.go:117] "RemoveContainer" containerID="0ce63c7ad4fb139ab4564105fa1322b7292f8c7a0f12262d42f4fcab821aa1ec" Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.435142 4934 generic.go:334] "Generic (PLEG): container finished" podID="c0cd0d57-203f-4d71-b26e-0329526839ed" containerID="dcc5d8adc2a50ddeae4081e4d0e95c36c3d8895302463f75f298a80479bf813a" exitCode=0 Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.435184 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdrbf" event={"ID":"c0cd0d57-203f-4d71-b26e-0329526839ed","Type":"ContainerDied","Data":"dcc5d8adc2a50ddeae4081e4d0e95c36c3d8895302463f75f298a80479bf813a"} Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.443374 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdfn7\" (UniqueName: \"kubernetes.io/projected/e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f-kube-api-access-zdfn7\") pod \"e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f\" (UID: \"e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f\") " Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.443633 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f-must-gather-output\") pod \"e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f\" (UID: \"e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f\") " Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.462695 4934 scope.go:117] "RemoveContainer" containerID="44b230a08f0db8ff213b6452d3159f62e7835d4f9033dfd4649d72ae1396c7d9" Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.464477 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f-kube-api-access-zdfn7" (OuterVolumeSpecName: "kube-api-access-zdfn7") pod "e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f" (UID: "e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f"). InnerVolumeSpecName "kube-api-access-zdfn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.548577 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdfn7\" (UniqueName: \"kubernetes.io/projected/e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f-kube-api-access-zdfn7\") on node \"crc\" DevicePath \"\"" Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.550486 4934 scope.go:117] "RemoveContainer" containerID="0ce63c7ad4fb139ab4564105fa1322b7292f8c7a0f12262d42f4fcab821aa1ec" Dec 27 09:28:46 crc kubenswrapper[4934]: E1227 09:28:46.550920 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ce63c7ad4fb139ab4564105fa1322b7292f8c7a0f12262d42f4fcab821aa1ec\": container with ID starting with 0ce63c7ad4fb139ab4564105fa1322b7292f8c7a0f12262d42f4fcab821aa1ec not found: ID does not exist" containerID="0ce63c7ad4fb139ab4564105fa1322b7292f8c7a0f12262d42f4fcab821aa1ec" Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.550963 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ce63c7ad4fb139ab4564105fa1322b7292f8c7a0f12262d42f4fcab821aa1ec"} err="failed to get container status \"0ce63c7ad4fb139ab4564105fa1322b7292f8c7a0f12262d42f4fcab821aa1ec\": rpc error: code = NotFound desc = could not find container \"0ce63c7ad4fb139ab4564105fa1322b7292f8c7a0f12262d42f4fcab821aa1ec\": container with ID starting with 0ce63c7ad4fb139ab4564105fa1322b7292f8c7a0f12262d42f4fcab821aa1ec not found: ID does not exist" Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.550988 4934 scope.go:117] "RemoveContainer" containerID="44b230a08f0db8ff213b6452d3159f62e7835d4f9033dfd4649d72ae1396c7d9" Dec 27 09:28:46 crc kubenswrapper[4934]: E1227 09:28:46.552757 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44b230a08f0db8ff213b6452d3159f62e7835d4f9033dfd4649d72ae1396c7d9\": container with ID starting with 44b230a08f0db8ff213b6452d3159f62e7835d4f9033dfd4649d72ae1396c7d9 not found: ID does not exist" containerID="44b230a08f0db8ff213b6452d3159f62e7835d4f9033dfd4649d72ae1396c7d9" Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.552806 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44b230a08f0db8ff213b6452d3159f62e7835d4f9033dfd4649d72ae1396c7d9"} err="failed to get container status \"44b230a08f0db8ff213b6452d3159f62e7835d4f9033dfd4649d72ae1396c7d9\": rpc error: code = NotFound desc = could not find container \"44b230a08f0db8ff213b6452d3159f62e7835d4f9033dfd4649d72ae1396c7d9\": container with ID starting with 44b230a08f0db8ff213b6452d3159f62e7835d4f9033dfd4649d72ae1396c7d9 not found: ID does not exist" Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.657715 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f" (UID: "e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.661611 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gdrbf" Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.755142 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0cd0d57-203f-4d71-b26e-0329526839ed-utilities\") pod \"c0cd0d57-203f-4d71-b26e-0329526839ed\" (UID: \"c0cd0d57-203f-4d71-b26e-0329526839ed\") " Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.755469 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0cd0d57-203f-4d71-b26e-0329526839ed-catalog-content\") pod \"c0cd0d57-203f-4d71-b26e-0329526839ed\" (UID: \"c0cd0d57-203f-4d71-b26e-0329526839ed\") " Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.755610 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xprhm\" (UniqueName: \"kubernetes.io/projected/c0cd0d57-203f-4d71-b26e-0329526839ed-kube-api-access-xprhm\") pod \"c0cd0d57-203f-4d71-b26e-0329526839ed\" (UID: \"c0cd0d57-203f-4d71-b26e-0329526839ed\") " Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.755952 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0cd0d57-203f-4d71-b26e-0329526839ed-utilities" (OuterVolumeSpecName: "utilities") pod "c0cd0d57-203f-4d71-b26e-0329526839ed" (UID: "c0cd0d57-203f-4d71-b26e-0329526839ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.757339 4934 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.757367 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0cd0d57-203f-4d71-b26e-0329526839ed-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.760004 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0cd0d57-203f-4d71-b26e-0329526839ed-kube-api-access-xprhm" (OuterVolumeSpecName: "kube-api-access-xprhm") pod "c0cd0d57-203f-4d71-b26e-0329526839ed" (UID: "c0cd0d57-203f-4d71-b26e-0329526839ed"). InnerVolumeSpecName "kube-api-access-xprhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.859958 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xprhm\" (UniqueName: \"kubernetes.io/projected/c0cd0d57-203f-4d71-b26e-0329526839ed-kube-api-access-xprhm\") on node \"crc\" DevicePath \"\"" Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.904622 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0cd0d57-203f-4d71-b26e-0329526839ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c0cd0d57-203f-4d71-b26e-0329526839ed" (UID: "c0cd0d57-203f-4d71-b26e-0329526839ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:28:46 crc kubenswrapper[4934]: I1227 09:28:46.962683 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0cd0d57-203f-4d71-b26e-0329526839ed-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 09:28:47 crc kubenswrapper[4934]: I1227 09:28:47.455037 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdrbf" event={"ID":"c0cd0d57-203f-4d71-b26e-0329526839ed","Type":"ContainerDied","Data":"de588c56b5238ca29b34271584e4e34f893d117f1835ce51028d9724f9759228"} Dec 27 09:28:47 crc kubenswrapper[4934]: I1227 09:28:47.455121 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gdrbf" Dec 27 09:28:47 crc kubenswrapper[4934]: I1227 09:28:47.456242 4934 scope.go:117] "RemoveContainer" containerID="dcc5d8adc2a50ddeae4081e4d0e95c36c3d8895302463f75f298a80479bf813a" Dec 27 09:28:47 crc kubenswrapper[4934]: I1227 09:28:47.493657 4934 scope.go:117] "RemoveContainer" containerID="4379488e8e9da6b76eab50eaf15124dee97387b1acf342547f19840e667b01f1" Dec 27 09:28:47 crc kubenswrapper[4934]: I1227 09:28:47.510710 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f" path="/var/lib/kubelet/pods/e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f/volumes" Dec 27 09:28:47 crc kubenswrapper[4934]: I1227 09:28:47.543171 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gdrbf"] Dec 27 09:28:47 crc kubenswrapper[4934]: I1227 09:28:47.565054 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gdrbf"] Dec 27 09:28:47 crc kubenswrapper[4934]: I1227 09:28:47.569431 4934 scope.go:117] "RemoveContainer" containerID="2060c39deb3b89715c8c86c3808846efcbaed4abc26dfc1f9aab8ea80d0a65fa" Dec 27 09:28:49 crc kubenswrapper[4934]: I1227 09:28:49.280985 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rztqh" Dec 27 09:28:49 crc kubenswrapper[4934]: I1227 09:28:49.281538 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rztqh" Dec 27 09:28:49 crc kubenswrapper[4934]: I1227 09:28:49.363346 4934 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rztqh" Dec 27 09:28:49 crc kubenswrapper[4934]: I1227 09:28:49.493728 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0cd0d57-203f-4d71-b26e-0329526839ed" path="/var/lib/kubelet/pods/c0cd0d57-203f-4d71-b26e-0329526839ed/volumes" Dec 27 09:28:49 crc kubenswrapper[4934]: I1227 09:28:49.547890 4934 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rztqh" Dec 27 09:28:50 crc kubenswrapper[4934]: I1227 09:28:50.531414 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rztqh"] Dec 27 09:28:51 crc kubenswrapper[4934]: I1227 09:28:51.507009 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rztqh" podUID="a602e20a-d41c-490c-af13-3fd0bffd4519" containerName="registry-server" containerID="cri-o://4d9e1a8088a7e48cc784ca16a921024ec63628f3d00492d287578808bd6713d1" gracePeriod=2 Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.092271 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rztqh" Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.199312 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a602e20a-d41c-490c-af13-3fd0bffd4519-utilities\") pod \"a602e20a-d41c-490c-af13-3fd0bffd4519\" (UID: \"a602e20a-d41c-490c-af13-3fd0bffd4519\") " Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.199407 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a602e20a-d41c-490c-af13-3fd0bffd4519-catalog-content\") pod \"a602e20a-d41c-490c-af13-3fd0bffd4519\" (UID: \"a602e20a-d41c-490c-af13-3fd0bffd4519\") " Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.199466 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qdnr\" (UniqueName: \"kubernetes.io/projected/a602e20a-d41c-490c-af13-3fd0bffd4519-kube-api-access-8qdnr\") pod \"a602e20a-d41c-490c-af13-3fd0bffd4519\" (UID: \"a602e20a-d41c-490c-af13-3fd0bffd4519\") " Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.200779 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a602e20a-d41c-490c-af13-3fd0bffd4519-utilities" (OuterVolumeSpecName: "utilities") pod "a602e20a-d41c-490c-af13-3fd0bffd4519" (UID: "a602e20a-d41c-490c-af13-3fd0bffd4519"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.208241 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a602e20a-d41c-490c-af13-3fd0bffd4519-kube-api-access-8qdnr" (OuterVolumeSpecName: "kube-api-access-8qdnr") pod "a602e20a-d41c-490c-af13-3fd0bffd4519" (UID: "a602e20a-d41c-490c-af13-3fd0bffd4519"). InnerVolumeSpecName "kube-api-access-8qdnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.226435 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a602e20a-d41c-490c-af13-3fd0bffd4519-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a602e20a-d41c-490c-af13-3fd0bffd4519" (UID: "a602e20a-d41c-490c-af13-3fd0bffd4519"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.302694 4934 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a602e20a-d41c-490c-af13-3fd0bffd4519-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.302750 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qdnr\" (UniqueName: \"kubernetes.io/projected/a602e20a-d41c-490c-af13-3fd0bffd4519-kube-api-access-8qdnr\") on node \"crc\" DevicePath \"\"" Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.302763 4934 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a602e20a-d41c-490c-af13-3fd0bffd4519-utilities\") on node \"crc\" DevicePath \"\"" Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.522314 4934 generic.go:334] "Generic (PLEG): container finished" podID="a602e20a-d41c-490c-af13-3fd0bffd4519" containerID="4d9e1a8088a7e48cc784ca16a921024ec63628f3d00492d287578808bd6713d1" exitCode=0 Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.522355 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rztqh" Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.522403 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rztqh" event={"ID":"a602e20a-d41c-490c-af13-3fd0bffd4519","Type":"ContainerDied","Data":"4d9e1a8088a7e48cc784ca16a921024ec63628f3d00492d287578808bd6713d1"} Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.522468 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rztqh" event={"ID":"a602e20a-d41c-490c-af13-3fd0bffd4519","Type":"ContainerDied","Data":"cdbb8f789ade948bb18a641ac14bdc13e6be3640393506bbecfba26b586f1297"} Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.522490 4934 scope.go:117] "RemoveContainer" containerID="4d9e1a8088a7e48cc784ca16a921024ec63628f3d00492d287578808bd6713d1" Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.554705 4934 scope.go:117] "RemoveContainer" containerID="8673421efe4027673d3c7a1ce8b3bfc4e08828d32bf5188b02f8f232c6dbe67b" Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.565620 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rztqh"] Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.575646 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rztqh"] Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.582661 4934 scope.go:117] "RemoveContainer" containerID="ed84b2b9f94ee71e5f55702f2589e1a3354c493288bd539faf2c0f909e37ce6e" Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.645111 4934 scope.go:117] "RemoveContainer" containerID="4d9e1a8088a7e48cc784ca16a921024ec63628f3d00492d287578808bd6713d1" Dec 27 09:28:52 crc kubenswrapper[4934]: E1227 09:28:52.645542 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d9e1a8088a7e48cc784ca16a921024ec63628f3d00492d287578808bd6713d1\": container with ID starting with 4d9e1a8088a7e48cc784ca16a921024ec63628f3d00492d287578808bd6713d1 not found: ID does not exist" containerID="4d9e1a8088a7e48cc784ca16a921024ec63628f3d00492d287578808bd6713d1" Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.645597 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d9e1a8088a7e48cc784ca16a921024ec63628f3d00492d287578808bd6713d1"} err="failed to get container status \"4d9e1a8088a7e48cc784ca16a921024ec63628f3d00492d287578808bd6713d1\": rpc error: code = NotFound desc = could not find container \"4d9e1a8088a7e48cc784ca16a921024ec63628f3d00492d287578808bd6713d1\": container with ID starting with 4d9e1a8088a7e48cc784ca16a921024ec63628f3d00492d287578808bd6713d1 not found: ID does not exist" Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.645623 4934 scope.go:117] "RemoveContainer" containerID="8673421efe4027673d3c7a1ce8b3bfc4e08828d32bf5188b02f8f232c6dbe67b" Dec 27 09:28:52 crc kubenswrapper[4934]: E1227 09:28:52.645901 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8673421efe4027673d3c7a1ce8b3bfc4e08828d32bf5188b02f8f232c6dbe67b\": container with ID starting with 8673421efe4027673d3c7a1ce8b3bfc4e08828d32bf5188b02f8f232c6dbe67b not found: ID does not exist" containerID="8673421efe4027673d3c7a1ce8b3bfc4e08828d32bf5188b02f8f232c6dbe67b" Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.645938 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8673421efe4027673d3c7a1ce8b3bfc4e08828d32bf5188b02f8f232c6dbe67b"} err="failed to get container status \"8673421efe4027673d3c7a1ce8b3bfc4e08828d32bf5188b02f8f232c6dbe67b\": rpc error: code = NotFound desc = could not find container \"8673421efe4027673d3c7a1ce8b3bfc4e08828d32bf5188b02f8f232c6dbe67b\": container with ID starting with 8673421efe4027673d3c7a1ce8b3bfc4e08828d32bf5188b02f8f232c6dbe67b not found: ID does not exist" Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.645959 4934 scope.go:117] "RemoveContainer" containerID="ed84b2b9f94ee71e5f55702f2589e1a3354c493288bd539faf2c0f909e37ce6e" Dec 27 09:28:52 crc kubenswrapper[4934]: E1227 09:28:52.646492 4934 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed84b2b9f94ee71e5f55702f2589e1a3354c493288bd539faf2c0f909e37ce6e\": container with ID starting with ed84b2b9f94ee71e5f55702f2589e1a3354c493288bd539faf2c0f909e37ce6e not found: ID does not exist" containerID="ed84b2b9f94ee71e5f55702f2589e1a3354c493288bd539faf2c0f909e37ce6e" Dec 27 09:28:52 crc kubenswrapper[4934]: I1227 09:28:52.646555 4934 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed84b2b9f94ee71e5f55702f2589e1a3354c493288bd539faf2c0f909e37ce6e"} err="failed to get container status \"ed84b2b9f94ee71e5f55702f2589e1a3354c493288bd539faf2c0f909e37ce6e\": rpc error: code = NotFound desc = could not find container \"ed84b2b9f94ee71e5f55702f2589e1a3354c493288bd539faf2c0f909e37ce6e\": container with ID starting with ed84b2b9f94ee71e5f55702f2589e1a3354c493288bd539faf2c0f909e37ce6e not found: ID does not exist" Dec 27 09:28:53 crc kubenswrapper[4934]: I1227 09:28:53.483160 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a602e20a-d41c-490c-af13-3fd0bffd4519" path="/var/lib/kubelet/pods/a602e20a-d41c-490c-af13-3fd0bffd4519/volumes" Dec 27 09:29:15 crc kubenswrapper[4934]: I1227 09:29:15.330402 4934 patch_prober.go:28] interesting pod/machine-config-daemon-w9j4r container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 27 09:29:15 crc kubenswrapper[4934]: I1227 09:29:15.331327 4934 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 27 09:29:15 crc kubenswrapper[4934]: I1227 09:29:15.331446 4934 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" Dec 27 09:29:15 crc kubenswrapper[4934]: I1227 09:29:15.332787 4934 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5"} pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 27 09:29:15 crc kubenswrapper[4934]: I1227 09:29:15.332895 4934 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerName="machine-config-daemon" containerID="cri-o://9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" gracePeriod=600 Dec 27 09:29:15 crc kubenswrapper[4934]: E1227 09:29:15.459229 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:29:15 crc kubenswrapper[4934]: I1227 09:29:15.838190 4934 generic.go:334] "Generic (PLEG): container finished" podID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" exitCode=0 Dec 27 09:29:15 crc kubenswrapper[4934]: I1227 09:29:15.838249 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" event={"ID":"a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7","Type":"ContainerDied","Data":"9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5"} Dec 27 09:29:15 crc kubenswrapper[4934]: I1227 09:29:15.838288 4934 scope.go:117] "RemoveContainer" containerID="f3e74ce13c29f6896d91ffd6e9150b3e5f47b60d6ed52c857f27a4eb1e7598d0" Dec 27 09:29:15 crc kubenswrapper[4934]: I1227 09:29:15.839427 4934 scope.go:117] "RemoveContainer" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" Dec 27 09:29:15 crc kubenswrapper[4934]: E1227 09:29:15.840039 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:29:30 crc kubenswrapper[4934]: I1227 09:29:30.469039 4934 scope.go:117] "RemoveContainer" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" Dec 27 09:29:30 crc kubenswrapper[4934]: E1227 09:29:30.470667 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:29:31 crc kubenswrapper[4934]: I1227 09:29:31.756433 4934 scope.go:117] "RemoveContainer" containerID="aa396c27848324ad149152cf42494d4e1b146e9be00c3264689f041563269213" Dec 27 09:29:41 crc kubenswrapper[4934]: I1227 09:29:41.468384 4934 scope.go:117] "RemoveContainer" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" Dec 27 09:29:41 crc kubenswrapper[4934]: E1227 09:29:41.469613 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:29:55 crc kubenswrapper[4934]: I1227 09:29:55.467892 4934 scope.go:117] "RemoveContainer" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" Dec 27 09:29:55 crc kubenswrapper[4934]: E1227 09:29:55.468683 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.178041 4934 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447130-pmq8q"] Dec 27 09:30:00 crc kubenswrapper[4934]: E1227 09:30:00.179811 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0cd0d57-203f-4d71-b26e-0329526839ed" containerName="extract-utilities" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.179846 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0cd0d57-203f-4d71-b26e-0329526839ed" containerName="extract-utilities" Dec 27 09:30:00 crc kubenswrapper[4934]: E1227 09:30:00.179876 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0cd0d57-203f-4d71-b26e-0329526839ed" containerName="extract-content" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.179886 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0cd0d57-203f-4d71-b26e-0329526839ed" containerName="extract-content" Dec 27 09:30:00 crc kubenswrapper[4934]: E1227 09:30:00.179902 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a602e20a-d41c-490c-af13-3fd0bffd4519" containerName="extract-utilities" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.179912 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a602e20a-d41c-490c-af13-3fd0bffd4519" containerName="extract-utilities" Dec 27 09:30:00 crc kubenswrapper[4934]: E1227 09:30:00.179935 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f" containerName="gather" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.179943 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f" containerName="gather" Dec 27 09:30:00 crc kubenswrapper[4934]: E1227 09:30:00.179965 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0cd0d57-203f-4d71-b26e-0329526839ed" containerName="registry-server" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.179972 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0cd0d57-203f-4d71-b26e-0329526839ed" containerName="registry-server" Dec 27 09:30:00 crc kubenswrapper[4934]: E1227 09:30:00.179987 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a602e20a-d41c-490c-af13-3fd0bffd4519" containerName="registry-server" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.179995 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a602e20a-d41c-490c-af13-3fd0bffd4519" containerName="registry-server" Dec 27 09:30:00 crc kubenswrapper[4934]: E1227 09:30:00.180025 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a602e20a-d41c-490c-af13-3fd0bffd4519" containerName="extract-content" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.180034 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="a602e20a-d41c-490c-af13-3fd0bffd4519" containerName="extract-content" Dec 27 09:30:00 crc kubenswrapper[4934]: E1227 09:30:00.180055 4934 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f" containerName="copy" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.180062 4934 state_mem.go:107] "Deleted CPUSet assignment" podUID="e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f" containerName="copy" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.180367 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="a602e20a-d41c-490c-af13-3fd0bffd4519" containerName="registry-server" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.180399 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f" containerName="gather" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.180421 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0cd0d57-203f-4d71-b26e-0329526839ed" containerName="registry-server" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.180455 4934 memory_manager.go:354] "RemoveStaleState removing state" podUID="e073c76c-2fff-4f2e-9bd2-5bd5932c5b8f" containerName="copy" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.181715 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447130-pmq8q" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.185027 4934 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.185526 4934 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.212650 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447130-pmq8q"] Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.279871 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c4b6d3f0-fafc-4cb6-a016-55bdd5d64883-secret-volume\") pod \"collect-profiles-29447130-pmq8q\" (UID: \"c4b6d3f0-fafc-4cb6-a016-55bdd5d64883\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447130-pmq8q" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.279910 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4b6d3f0-fafc-4cb6-a016-55bdd5d64883-config-volume\") pod \"collect-profiles-29447130-pmq8q\" (UID: \"c4b6d3f0-fafc-4cb6-a016-55bdd5d64883\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447130-pmq8q" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.279940 4934 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7vrg\" (UniqueName: \"kubernetes.io/projected/c4b6d3f0-fafc-4cb6-a016-55bdd5d64883-kube-api-access-b7vrg\") pod \"collect-profiles-29447130-pmq8q\" (UID: \"c4b6d3f0-fafc-4cb6-a016-55bdd5d64883\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447130-pmq8q" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.382809 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c4b6d3f0-fafc-4cb6-a016-55bdd5d64883-secret-volume\") pod \"collect-profiles-29447130-pmq8q\" (UID: \"c4b6d3f0-fafc-4cb6-a016-55bdd5d64883\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447130-pmq8q" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.382867 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4b6d3f0-fafc-4cb6-a016-55bdd5d64883-config-volume\") pod \"collect-profiles-29447130-pmq8q\" (UID: \"c4b6d3f0-fafc-4cb6-a016-55bdd5d64883\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447130-pmq8q" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.382904 4934 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7vrg\" (UniqueName: \"kubernetes.io/projected/c4b6d3f0-fafc-4cb6-a016-55bdd5d64883-kube-api-access-b7vrg\") pod \"collect-profiles-29447130-pmq8q\" (UID: \"c4b6d3f0-fafc-4cb6-a016-55bdd5d64883\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447130-pmq8q" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.383761 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4b6d3f0-fafc-4cb6-a016-55bdd5d64883-config-volume\") pod \"collect-profiles-29447130-pmq8q\" (UID: \"c4b6d3f0-fafc-4cb6-a016-55bdd5d64883\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447130-pmq8q" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.389799 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c4b6d3f0-fafc-4cb6-a016-55bdd5d64883-secret-volume\") pod \"collect-profiles-29447130-pmq8q\" (UID: \"c4b6d3f0-fafc-4cb6-a016-55bdd5d64883\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447130-pmq8q" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.406800 4934 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7vrg\" (UniqueName: \"kubernetes.io/projected/c4b6d3f0-fafc-4cb6-a016-55bdd5d64883-kube-api-access-b7vrg\") pod \"collect-profiles-29447130-pmq8q\" (UID: \"c4b6d3f0-fafc-4cb6-a016-55bdd5d64883\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29447130-pmq8q" Dec 27 09:30:00 crc kubenswrapper[4934]: I1227 09:30:00.511434 4934 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447130-pmq8q" Dec 27 09:30:01 crc kubenswrapper[4934]: I1227 09:30:01.017052 4934 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447130-pmq8q"] Dec 27 09:30:01 crc kubenswrapper[4934]: I1227 09:30:01.437001 4934 generic.go:334] "Generic (PLEG): container finished" podID="c4b6d3f0-fafc-4cb6-a016-55bdd5d64883" containerID="797717bf8662a9c05974b32432c076e2227c021d225c063e49c10b6ab641c053" exitCode=0 Dec 27 09:30:01 crc kubenswrapper[4934]: I1227 09:30:01.437044 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447130-pmq8q" event={"ID":"c4b6d3f0-fafc-4cb6-a016-55bdd5d64883","Type":"ContainerDied","Data":"797717bf8662a9c05974b32432c076e2227c021d225c063e49c10b6ab641c053"} Dec 27 09:30:01 crc kubenswrapper[4934]: I1227 09:30:01.437071 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447130-pmq8q" event={"ID":"c4b6d3f0-fafc-4cb6-a016-55bdd5d64883","Type":"ContainerStarted","Data":"5318a54f3b607c37a6a99e9b93347cbccc91d3979118e4da7bf296513a6c5054"} Dec 27 09:30:02 crc kubenswrapper[4934]: I1227 09:30:02.895549 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447130-pmq8q" Dec 27 09:30:03 crc kubenswrapper[4934]: I1227 09:30:03.054064 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7vrg\" (UniqueName: \"kubernetes.io/projected/c4b6d3f0-fafc-4cb6-a016-55bdd5d64883-kube-api-access-b7vrg\") pod \"c4b6d3f0-fafc-4cb6-a016-55bdd5d64883\" (UID: \"c4b6d3f0-fafc-4cb6-a016-55bdd5d64883\") " Dec 27 09:30:03 crc kubenswrapper[4934]: I1227 09:30:03.054840 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4b6d3f0-fafc-4cb6-a016-55bdd5d64883-config-volume\") pod \"c4b6d3f0-fafc-4cb6-a016-55bdd5d64883\" (UID: \"c4b6d3f0-fafc-4cb6-a016-55bdd5d64883\") " Dec 27 09:30:03 crc kubenswrapper[4934]: I1227 09:30:03.055073 4934 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c4b6d3f0-fafc-4cb6-a016-55bdd5d64883-secret-volume\") pod \"c4b6d3f0-fafc-4cb6-a016-55bdd5d64883\" (UID: \"c4b6d3f0-fafc-4cb6-a016-55bdd5d64883\") " Dec 27 09:30:03 crc kubenswrapper[4934]: I1227 09:30:03.055747 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4b6d3f0-fafc-4cb6-a016-55bdd5d64883-config-volume" (OuterVolumeSpecName: "config-volume") pod "c4b6d3f0-fafc-4cb6-a016-55bdd5d64883" (UID: "c4b6d3f0-fafc-4cb6-a016-55bdd5d64883"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 27 09:30:03 crc kubenswrapper[4934]: I1227 09:30:03.059846 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4b6d3f0-fafc-4cb6-a016-55bdd5d64883-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c4b6d3f0-fafc-4cb6-a016-55bdd5d64883" (UID: "c4b6d3f0-fafc-4cb6-a016-55bdd5d64883"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 27 09:30:03 crc kubenswrapper[4934]: I1227 09:30:03.059884 4934 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4b6d3f0-fafc-4cb6-a016-55bdd5d64883-kube-api-access-b7vrg" (OuterVolumeSpecName: "kube-api-access-b7vrg") pod "c4b6d3f0-fafc-4cb6-a016-55bdd5d64883" (UID: "c4b6d3f0-fafc-4cb6-a016-55bdd5d64883"). InnerVolumeSpecName "kube-api-access-b7vrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 27 09:30:03 crc kubenswrapper[4934]: I1227 09:30:03.158002 4934 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4b6d3f0-fafc-4cb6-a016-55bdd5d64883-config-volume\") on node \"crc\" DevicePath \"\"" Dec 27 09:30:03 crc kubenswrapper[4934]: I1227 09:30:03.158055 4934 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c4b6d3f0-fafc-4cb6-a016-55bdd5d64883-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 27 09:30:03 crc kubenswrapper[4934]: I1227 09:30:03.158072 4934 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7vrg\" (UniqueName: \"kubernetes.io/projected/c4b6d3f0-fafc-4cb6-a016-55bdd5d64883-kube-api-access-b7vrg\") on node \"crc\" DevicePath \"\"" Dec 27 09:30:03 crc kubenswrapper[4934]: I1227 09:30:03.461894 4934 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29447130-pmq8q" event={"ID":"c4b6d3f0-fafc-4cb6-a016-55bdd5d64883","Type":"ContainerDied","Data":"5318a54f3b607c37a6a99e9b93347cbccc91d3979118e4da7bf296513a6c5054"} Dec 27 09:30:03 crc kubenswrapper[4934]: I1227 09:30:03.461949 4934 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5318a54f3b607c37a6a99e9b93347cbccc91d3979118e4da7bf296513a6c5054" Dec 27 09:30:03 crc kubenswrapper[4934]: I1227 09:30:03.462117 4934 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29447130-pmq8q" Dec 27 09:30:03 crc kubenswrapper[4934]: I1227 09:30:03.966712 4934 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447085-mnn49"] Dec 27 09:30:03 crc kubenswrapper[4934]: I1227 09:30:03.977020 4934 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29447085-mnn49"] Dec 27 09:30:05 crc kubenswrapper[4934]: I1227 09:30:05.493769 4934 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="868f30f9-ef52-458b-a35c-193269bfba46" path="/var/lib/kubelet/pods/868f30f9-ef52-458b-a35c-193269bfba46/volumes" Dec 27 09:30:07 crc kubenswrapper[4934]: I1227 09:30:07.469473 4934 scope.go:117] "RemoveContainer" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" Dec 27 09:30:07 crc kubenswrapper[4934]: E1227 09:30:07.470681 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:30:21 crc kubenswrapper[4934]: I1227 09:30:21.467773 4934 scope.go:117] "RemoveContainer" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" Dec 27 09:30:21 crc kubenswrapper[4934]: E1227 09:30:21.469067 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:30:31 crc kubenswrapper[4934]: I1227 09:30:31.890479 4934 scope.go:117] "RemoveContainer" containerID="82d5e1c7a3c2a693188e5e1754b115498b98c16a15a8af6a3bc44a1d55a7f3ef" Dec 27 09:30:34 crc kubenswrapper[4934]: I1227 09:30:34.480501 4934 scope.go:117] "RemoveContainer" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" Dec 27 09:30:34 crc kubenswrapper[4934]: E1227 09:30:34.481927 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:30:48 crc kubenswrapper[4934]: I1227 09:30:48.469306 4934 scope.go:117] "RemoveContainer" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" Dec 27 09:30:48 crc kubenswrapper[4934]: E1227 09:30:48.470561 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:31:02 crc kubenswrapper[4934]: I1227 09:31:02.469341 4934 scope.go:117] "RemoveContainer" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" Dec 27 09:31:02 crc kubenswrapper[4934]: E1227 09:31:02.470396 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:31:13 crc kubenswrapper[4934]: I1227 09:31:13.469153 4934 scope.go:117] "RemoveContainer" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" Dec 27 09:31:13 crc kubenswrapper[4934]: E1227 09:31:13.470630 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:31:25 crc kubenswrapper[4934]: I1227 09:31:25.468677 4934 scope.go:117] "RemoveContainer" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" Dec 27 09:31:25 crc kubenswrapper[4934]: E1227 09:31:25.469934 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:31:38 crc kubenswrapper[4934]: I1227 09:31:38.467527 4934 scope.go:117] "RemoveContainer" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" Dec 27 09:31:38 crc kubenswrapper[4934]: E1227 09:31:38.468243 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:31:53 crc kubenswrapper[4934]: I1227 09:31:53.467846 4934 scope.go:117] "RemoveContainer" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" Dec 27 09:31:53 crc kubenswrapper[4934]: E1227 09:31:53.468898 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:32:08 crc kubenswrapper[4934]: I1227 09:32:08.468335 4934 scope.go:117] "RemoveContainer" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" Dec 27 09:32:08 crc kubenswrapper[4934]: E1227 09:32:08.470018 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:32:22 crc kubenswrapper[4934]: I1227 09:32:22.467597 4934 scope.go:117] "RemoveContainer" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" Dec 27 09:32:22 crc kubenswrapper[4934]: E1227 09:32:22.468847 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:32:37 crc kubenswrapper[4934]: I1227 09:32:37.467937 4934 scope.go:117] "RemoveContainer" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" Dec 27 09:32:37 crc kubenswrapper[4934]: E1227 09:32:37.468648 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:32:51 crc kubenswrapper[4934]: I1227 09:32:51.467408 4934 scope.go:117] "RemoveContainer" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" Dec 27 09:32:51 crc kubenswrapper[4934]: E1227 09:32:51.468569 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:33:06 crc kubenswrapper[4934]: I1227 09:33:06.467691 4934 scope.go:117] "RemoveContainer" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" Dec 27 09:33:06 crc kubenswrapper[4934]: E1227 09:33:06.468688 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:33:17 crc kubenswrapper[4934]: I1227 09:33:17.468120 4934 scope.go:117] "RemoveContainer" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" Dec 27 09:33:17 crc kubenswrapper[4934]: E1227 09:33:17.469068 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:33:28 crc kubenswrapper[4934]: I1227 09:33:28.468559 4934 scope.go:117] "RemoveContainer" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" Dec 27 09:33:28 crc kubenswrapper[4934]: E1227 09:33:28.469557 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" Dec 27 09:33:43 crc kubenswrapper[4934]: I1227 09:33:43.468478 4934 scope.go:117] "RemoveContainer" containerID="9dd9fed1c4fbe9d77f7759f4d4dea5c00e62112e9e7ee41cb62c258aa101cdf5" Dec 27 09:33:43 crc kubenswrapper[4934]: E1227 09:33:43.469437 4934 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-w9j4r_openshift-machine-config-operator(a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7)\"" pod="openshift-machine-config-operator/machine-config-daemon-w9j4r" podUID="a8f0a3d2-4eb3-46cc-91a0-ba934794e0d7" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515123724005024444 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015123724005017361 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015123706375016516 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015123706376015467 5ustar corecore